Ibm Spss Statistics Excellent Guide
Ibm Spss Statistics Excellent Guide
EXCELLENT GUIDE
1
Copyright © 2020 by Peter James Kpolovie.
ISBN: eBook
Print
Except as permitted under the U.S. Copyright Law, no part of this book
may be reprinted by any means, graphic, electronic, or mechanical,
including photocopying, recording, taping or by any information storage
retrieval system without the written permission of the author, excluding the
case of brief quotations embodied in critical articles and reviews with due
referencing.
2
Table of Contents
PREFACE
Pedagogical features
Final benefit
3
Printing of Data File
Printing of Syntax File
RETRIEVAL OF FILES
Retrieval of Data File
Retrieving Syntax File
Retrieval of Output File
EXITING IBM SPSS STATISTICS
4
FREQUENCIES SPSS SYNTAX FOR DESCRIPYIVE STATISTICS
Syntax 2.4 Chapter 2 Syntax 4.sps
Output 2.8 Chapter 2 Output 8.spv
5
Output 3.4 Chapter 3 Output 4.spv
6
EFFECT SIZE, d, IN PAIRED-SAMPLES t TEST
SPSS SYNTAX FOR PAIRED-SAMPLES t TEST ON SPOUSES’ SEA
Syntax 5.1 Chapter 5 Syntax 1.sps
Output 5.2 Chapter 5 Output 2.spv
SELF-ACTUALIZATION IN 3-YEAR UNDERGRADUATE
PROGRAMMES
Research question
Alternate hypothesis
Null hypothesis
SPSS DIALOG BOXES SELECTION FOR PAIRED-SAMPLES t TEST
CASE 2
Output 5.3 Chapter 5 Output 3.spv
INTERPRETATION OF OUTPUT 5.3: SELF-ACTUALIZATION
EFFECT SIZE, d
SPSS SYNTAX FOR PAIRED-SAMPLES t TEST ON SELF-
ACTUALIZATION
Syntax 5.2 Chapter 5 Syntax 2.sps
Output 5.4 Chapter 5 Output 4.spv
7
EFFECT SIZE IN POST HOC MULTIPLE COMPARISONS
Effect Size, d, for 8W LED versus 9W CFL Bulbs
Effect Size, d, for 9W LED versus 40W Incandescent Bulbs
Effect Size, d, for 8W LED versus 40W Incandescent Bulbs
SUMMARY OF THE FINDINGS
SPSS SYNTAX METHOD OF ANOVA
Syntax 6.1 Chapter 6 Syntax 1.sps
Output 6.2 Chapter 6 Output 2.spv
INTERPRETATION OF OUTPUT 6.2: ONE-WAY ANOVA LUMINOUS
EFFICACY
UNIVARIATE SPSS DIALOG BOXES POINT AND CLICK APPROACH
TO ANOVA
Output 6.3 Chapter 6 Output 3
Univariate Analysis of Variance
Estimated Marginal Means
Post Hoc Tests
Homogeneous Subsets
Profile Plots
INTERPRETATION OF OUTPUT 6.3: UNIVARIATE ANOVA ON BULB
TYPES
Between-subjects Factors
Descriptive Statistics
Tests of Between-Subjects Effects
EFFECT SIZE IN UNIVARIATE ANOVA
Partial Eta Squared, ɳp2
EFFECT SIZE AS R SQUARED IN UNIVARAITE ANOVA
EFFECT SIZE COMPUTION AS PARTIAL ETA SQUARED
Estimated Marginal Means for Types of Bulb
Pairwise Comparisons
Univariate Tests
Post Hoc Tests on Types of Bulb Multiple Comparisons
Homogeneous Subsets
EFFECT SIZE FOR EACH PAIRWISE MEAN DIFFERENCE
Profile Plots for the Estimated Marginal Means of Luminous Efficacy
SUMMARY OF THE RESULTS
UNIVARIATE SPSS SYNTAX APPROACH FOR ANOVA
Syntax 6.2 Chapter 6 Syntax 2.sps
8
Output 6.4 Chapter 6 Output 4.spv
INTERPRETATION OF OUTPUT 6.4
9
Chapter 8 - ONE-WAY REPEATED MEASURES ANALYSIS OF
VARIANCE
Overview
EFFECTS OF NURSING INTERVENTIONS ON PATIENTS’
COMPREHENSIVE HEALTH
Research question
Alternate hypothesis
Null hypothesis
ONE-WAY REPEATED MEASURES ANOVA WITH SPSS DIALOG
BOXES SELECTION
FURTHER FOLLOW-UP PAIRWISE COMPARISONS TESTS
Output 8.1 Chapter 8 Output 1.spv
Further Pairwise Tests using Paired-Samples t Test
INTERPRETATION OF OUTPUT 8.1: EFFECTS OF NURSING
INTERVENTIONS
Within-Subjects Factors
Descriptive Statistics
Multivariate Tests
Mauchly’s Test of Sphericity
Tests of Within-Subjects Effects
Effect Size
Tests of Between-Subjects Effects and Tests of Within Subjects Contrast
Estimated Marginal Means for Interventions
Pairwise Comparisons
Multivariate Tests
Profile Plots
Paired Samples Statistics
Paired Samples Correlations
Paired Samples Test of Paired Differences
ONE-WAY REPEATED MEASURES ANOVA WITH SPSS SYNTAX
Syntax 8.1 Chapter 8 syntax 1.sps
Output 8.2 Chapter 8 Output 2.spv
Further Pairwise Tests with Paired-Samples t Test
INTERPRETATION OF OUTPUT 8.2
SCHOLARLY PERFORMANCE IN 5-YEAR DIRECT Ph.D.
ENGINEERING PROGRAM
Research question
10
Alternate Hypothesis
Null Hypothesis
ONE-WAY WITHIN-SUBJECTS ANOVA WITH SPSS DIALOG BOXES
SELECTION
FOLLOW-UP PAIRWISE COMPARISONS TESTS
Output 8.3 Chapter 8 Output 3.spv – Years of study and scholarly
performance
Pairwise Tests as Follow-up
INTERPRETATION OF THE OUTPUT 8.3 – INFLUENCE OF YEARS
OF STUDY
Within-Subjects Factors
Descriptive Statistics
Multivariate Tests
Mauchly’s Test of Sphericity
Tests of Within-Subjects Effects
Effect Size
Test of Between-Subjects Effects and Test of Wuthin-Subjects Contrast
Estimated Marginal Means for Year
Pairwise Comparisons
Multivariate Tests
Profile Plots
Paired Samples Statistics
Paired Samples Correlations
Paired Samples Test of Paired Differences
ONE-WAY WITHIN-SUBJECTS ANOVA WITH SPSS SYNTAX
Syntax 8.2 Chapter 8 Syntax 2.sps
Output 8.4 Chapter 8 Output 4.spv – Years of study and scholarly
performance
Pairwise Follow-up Tests
INTERPRETATION OF OUTPUT 8.4
COMPARISON OF THE TWO ONE-WAY REPEATED MEASURES
ANOVA SYNTAX
11
EFFECTS OF A NEW MEMORY DRUG AND LEARNING STRATEGY
ON LONG-TERM MEMORY
Research Questions
Alternate Hypotheses
Null Hypotheses
SPSS TWO-WAY REPEATED MEASURES ANOVA VIA DIALOG
BOXES POINT AND CLICK
Output 9.1 Chapter 9 Output 1.spv
INTERPRETATION OF OUTPUT 9.1 – MEMORY DRUG AND
LEARNING STRATEGY
Within-subjects Factors
Descriptive Statistics
Multivariate Tests
Mauchly’s Test of Sphericity
Tests of Within-Subjects Effects
Tests of Within-Subjects Contrasts
Tests of Between-Subjects Effects
Estimated Marginal Means for MemDrug Estimates
Estimated Marginal Means for MemDrug Pairwise Comparisons
Multivariate Tests
Estimated Marginal Means for LearnStrategy Estimates
Estimated Marginal Means for LearnStrategy Pairwise Comparisons
Multivariate Tests
Estimated Marginal Means for MemDrug*LearnStrategy
Profile Plots
SUMMARY OF THE TWO-WAY REPEATED MEASURES ANOVA
TWO-WAY REPEATED MEASURES ANOVA EXECUTION WITH
SPSS SYNTAX
Syntax 9.1 Two-Way Repeated Measures ANOVA Syntax
Output 9.2 Chapter 9 Output 2.spv – Memory Drug and Learning Strategy
INTERPRETATION OF OUTPUT 9.2
12
STAFF RECRUITMENT EXERCISE FOR A GIFTED CHILDREN’S
SCHOOL STEM PROGRAM
Research questions
Alternate Hypotheses
Null hypotheses
TWO-WAY MIXED (BETWEEN-WITHIN) ANOVA WITH SPSS
DIALOG BOXES SELECTION
MANOVA FOR FOLLOW-UP OF STEM * DISCIPLINE SIGNIFICANT
INTERACTION EFFECT ON SREP
Syntax 10.1 Chapter 10 Syntax 1.sps – MANOVA Syntax for follow-up
Output 10.1 Chapter 10 Output 1.spv
RESULTS OF MANOVA FOR INTERACTION FOLLOW-UP
INTERPRETATION OF OUTPUT 10.1 – EFFECTS OF STEM AND
DISCIPLINE
Within-Subjects Factors
Between-Subjects Factors
Descriptive Statistics
Multivariate Tests
Mauchly’s Test of Sphericity
Tests of Within-Subjects Effects
Tests of Within-Subjects Contrasts
Tests of Between-Subjects Effects
Estimated Marginal Means for Discipline
Pairwise Comparisons
Univariate Tests
Estimated Marginal Means for Stem
Pairwise Comparisons
Multivariate Tests
Estimated Marginal Means for Discipline * Stem
Post Hoc Tests for Discipline
Homogeneous Subsets for Srep
Profile Plots
INTERPRETING THE MANOVA INTERACTION FOLLOW-UP TESTS
Tests involving ‘MWITHIN STEM(1)’ Within-Subject Effect
Tests involving ‘MWITHIN STEM(2)’ Within-Subject Effect
Tests involving ‘MWITHIN STEM(3)’ Within-Subject Effect
Tests involving ‘MWITHIN STEM(4)’ Within-Subject Effect
13
Tests involving ‘STEM’ Within-Subject Effect
Tests involving ‘STEM’ Within-Subject Effect
Tests involving ‘STEM’ Within-Subject Effect
Tests of Between-Subjects Effects
Tests involving ‘STEM’ Within-Subject Effect
EFFECT .. MWITHIN DISCIPLINE(3) BY STEM
EFFECT .. MWITHIN DISCIPLINE(2) BY STEM
EFFECT .. MWITHIN DISCIPLINE(1) BY STEM
SUMMARY OF THE TWO-WAY MIXED (BETWEEN-WITHIN)
ANOVA OUTPUT
SPSS SYNTAX FOR TWO-WAY MIXED (BETWEEN-WITHIN)
ANOVA
Syntax 10.2 Chapter 10 Syntax 2.sps
Output 10.2 Chapter 10 Output 2.spv
RESULTS OF MANOVA FOR FOLLOW-UP TESTS
INTERPRETATION OF OUTPUT 10.2 – EFFECTS OF STEM AND
DISCIPLINE
14
SECTION B – ANOVA FOR COMPARISON OF THE F WITH THAT OF
ANCOVA
Output 11.1 Chapter 11 Output 1.spv – Effect of LumosityTS on recall
SECTION B OF OUTPUT 11.1 One-Way ANOVA
INTERPRETATION OF OUTPUT 11.1 – EFFECT OF LUMOSITY
TRAINING SCHEDULE
Univariate Analysis of Variance Between-Subjects Factors
Descriptive Statistics
Levene’s Test of Equality of Error Variances
Tests of Between-Subjects Effects
EFFECT SIZE IN THE ANCOVA ON NONSENSE SYLLABLES
RECALL
Estimated Marginal Means for LumosityTS NSPost-test
Pairwise Comparisons of NSPost-test
Univariate Tests
Profile Plots
SECTION B: ONE-WAY ANOVA AS A FOLLOW-UP TEST
Oneway Descriptives
Test of Homogeneity of Variances
ANOVA
Effect Size in the ANOVA
Post Hoc Tests Multiple Comparisons with Bonferroni
Means Plots
SUMMARY OF THE RESULTS
ANCOVA EXECUTION WITH SPSS SYNTAX
Syntax 11.1 Chapter 11 Syntax 1.sps – Syntax for performing ANCOVA
Output 11.2 Chapter 11 Output 2.sps – Effect of Lumosity Training
schedule on recall
SECTION B OF OUTPUT 11.1 One-Way ANOVA
INTERPRETATION OF OUTPUT 11.2
EFFECT OF LUMOSITY TRAINING SCHEDULE ON INTELLIGENCE
QUOTIENT (IQ)
Research questions
Alternate Hypothesis
Null Hypothesis
ANCOVA WITH SPSS DIALOG BOXES POINT AND CLICK
ONE-WAY ANOVA AS ANCOVA FOLLOW-UP
15
Output 11.3 Chapter 11 Output 3.spv – ANCOVA for effect of Lumosity
training on IQ
INTERPRETATION OF OUTPUT 11.3 – EFFECT OF LUMOSITY
TRAINING SCHEDULE ON IQ
Univariate Analysis of Variance Between-Subjects Factors
Descriptive Statistics
Levene’s Test of Equality of Error Variances
Tests of Between-Subjects Effects
EFFECT SIZE IN THE ANCOVA
Estimated Marginal Means for LumosityTS Post-test IQ
Pairwise Comparisons of Post-test IQ
Univariate Tests
Profile Plots
SECTION B: ONE-WAY ANOVA AS FOLLOW-UP TEST: Oneway
Descriptives
Test of Homogeneity of Variances
ANOVA
Post Hoc Tests Multiple Comparisons with Bonferroni
Means Plots
SUMMARY OF THE RESULTS
ANCOVA WITH IBM SPSS SYNTAX FOR LUMOSITYTS ON IQPOST
Syntax 11.2 Chapter 11 Syntax 2.sps
Output 11.4 Chapter 11 Output 4.spv
INTERPRETATION OF OUTPUT 11.4
16
Syntax 4.2 Chapter 4 Syntax 2.sps
Syntax 5.1 Chapter 5 Syntax 1.sps
Syntax 5.2 Chapter 5 Syntax 2.sps
Syntax 6.1 Chapter 6 Syntax 1.sps
Syntax 6.2 Chapter 6 Syntax 2.sps
Syntax 7.1 Chapter 7 Syntax 1.sps
Syntax 8.1 Chapter 8 syntax 1.sps
Syntax 8.2 Chapter 8 Syntax 2.sps
Syntax 9.1 Chapter 9 Syntax 1.sps
Syntax 10.1 Chapter 10 Syntax 1.sps
Syntax 10.2 Chapter 10 Syntax 2.sps
Syntax 11.1 Chapter 11 Syntax 1.sps
Syntax 11.2 Chapter 11 Syntax 2.sps
REFERENCES
17
Preface
It is the most practical tool for the actualization of the paradigm shift
from the use of a hand-held calculator and some unnecessarily complicated
and slow statistical software to the flawlessly swift application of IBM
SPSS Statistics in data analysis. Once a person acquires a copy, he can
independently apply SPSS Statistics accurately in analysing different sorts
of data, irrespective of the volume and complexity of the datasets. It is a
timeless ideal guide to all the categories of users as it makes no assumptions
of previous statistical knowledge and experience.
What if you can analyze different sorts of data correctly and expertly
interpret the output? Virtually in all fields of endeavour, some data could be
collected quantitatively and most easily analysed, using SPSS Statistics as
thoroughly presented in this textbook.
18
IBM SPSS Statistics is the most powerful and widely used software for
data analysis in the social and behavioral sciences, particularly, and in other
several different fields of endeavor. Practical statistical analysis serves as a
fantastic set of commitments that is necessarily required to be a scientist or
researcher in the contemporary world.
The scientific world is progressing decisively away from the manual
computation of data with a hand-held calculator that is tedious, daunting,
and limited in memory and capacity to the application of statistical software
for the swiftest data analysis with the maximum precision. The use of a
hand-held calculator once worked very well, but it could no longer cope
with the complexity and enormity of research data that the current and
subsequent centuries demand. Unfortunately, however, skillful use of the
software, SPSS Statistics, for data analysis is yet to be mastered by an
overwhelming majority of individuals, even those who frequently work
with quantitative and qualitative data. The book explicitly offers solutions
to problems of data management, analysis and interpretation of the output.
This book automatically guarantees each user an incomparably better,
faster, and more accurate ways of getting data for each statistical test
analysed with the use of IBM SPSS Statistics than the traditional method of
analysing the same data with a hand-held calculator or even with some
other statistical software that are slow, arduous, difficult and very
cumbersome to apply. The book does an excellent job at explaining the
latest SPSS Statistics user interface most pragmatically with complete step-
by-step screenshots and an unprecedented focus on interpreting the results.
Throughout the history of science, nothing gives the scientist greater
satisfaction and self-fulfilment than finding a more straightforward and
swifter technique of arriving at realistic solutions to a persistent problem.
To accurately analyse large complex dataset collected for a given research
work has consistently remained a problem to the investigator before the
actual problem that he has set out to investigate. This is due to the lack of
easy but complete guide on how to use the powerful software, SPSS
Statistics. With IBM SPSS Statistics Excellent Guide, the researcher can
with most exceptional effortlessness utilize SPSS and follow the unbroken
screenshots, excellently illustrated in an easy to use point-by-point guide to
analyse the dada personally impeccably.
An investigator who has a copy of the book will be saved from the
associated harm of painstakingly carrying out experimental observations
19
flawlessly and handing over the data to another person, often a technically
incompetent person, for analysis who may carefreely perform wrong
analysis for the investigator. With the book, each investigator can, with all
accuracy, perform his data analysis personally. Even if it is his first day of
using SPSS Statistics, he cannot make any mistake in the analysis as long as
he follows the exciting comprehensive, pragmatic cutting-edge step by step
guide demonstrated in the text.
The IBM SPSS Statistics Excellent Guide is in series that are aimed at
equipping each user with the revolutionary skills for accurate application of
SPSS statistical software in data analytics for value-added decision making
with descriptive, diagnostic, predictive, causal explanatory, and prescriptive
control of phenomena. In this first in the series, myriads of introductory and
advanced statistical approaches are treated explicitly that include
Descriptive statistics, Comparison of means, t Test, Analysis of Variance,
Univariate analysis, Repeated measures analysis, Analysis of Covariance,
and Three-Way Analysis of Variance.
Having the book shall equally be an invaluable asset for each person in
our digital world as it offers the latest data analytic skills, and ensures a
better understanding of others' research findings. This book is a must-have
resource for undergraduates and graduate students throughout the world
who are embarking on or about to execute their research projects, theses or
dissertations. Also, researchers, instructors, scientists, professionals, and
data-based decision-makers definitely need the book for automation of their
data analysis. For anyone who wants to earn a good living legitimately or
who wants to make additional income from data analysis, getting and using
IBM SPSS Statistics Excellent Guide series is the sure investment.
The IBM SPSS Statistics Excellent Guide will ever remain an invaluable
fundamental text and resource material for numerous courses. In the
countless institutions where courses such as 1) Descriptive statistics; 2)
Introduction to SPSS Statistics; 3) Data Analysis with Statistical Software;
4) Advanced Analytics; 5) Research Methods; 6) Basic Statistics; 7) ICT
and Research; 8) Univariate Analysis of Variance; 9) Data Analysis with
Computer; 10) Effect Size and Power Analysis; 11) SPSS Core Skill; 12)
SPSS Research Resources, and etc. are offered, this book most
appropriately fits in as the core textbook.
With reference to MSC2010 Mathematics Subjects Classification 2010,
courses that the students will fundamentally use this book suitably include
20
the following among others:
i. 62K86 – Design of experiments
ii. 62J15 – Paired and multiple comparisons
iii. 62H86 – Multiple analysis
iv. 62J10 – Analysis of variance and covariance
v. 97K40 – Descriptive statistics
vi. 62K15 – Factorial designs
vii. 62H15 – Hypothesis testing
viii. 97K80 – Applied statistics
ix. 62J12 – General and Generalized linear models
As long as research rules the world, IBM SPSS Statistics Excellent Guide
shall be in very high demand in all parts of the globe for the swiftest and
most precise data analysis in the Information and Communication
Technology-driven Age. A vast majority of academic degree programs
demand a research project, thesis or dissertation as a compulsory
requirement for graduation. Most of such research involves the collection of
quantified data that could best be analyzed with SPSS Statistics as
exceptionally presented in this book. Virtually all universities and colleges
offer some research-related courses that require data analysis, as
demonstrated in this book with SPSS Statistics. For all such courses, this
book is a wonderful supplementary resource.
Millions of university and college students (undergraduates and
graduates) annually worldwide who must execute quantitative or qualitative
data-based research as a precondition for successful completion of their
programs necessarily need to acquire the book for personal analysis of their
data and interpretation of the outputs. Researchers who must from time to
time collect data, analyse and disseminate their findings for contribution to
knowledge in line with the established principle of "publish or perish"
necessarily need to acquire the book to analyse, interpret and report their
investigations accurately.
IBM SPSS Statistics Excellent Guide is the superb book for
independently guiding a researcher to employ the right research method,
apply the best statistical technique to the observed data, execute the entire
analysis personally with all precision effortlessly, and report the findings.
The book is also structured to be an asset of inestimable value to
students who are running Open and Distance Learning programmes and
other online academic programmes as it is a totally illustrated hands-on
21
point-by-point excellent guide for SPSS Statistics application. It is written
to meet the needs of the very fast-growing millions of students in online
academic programmes and in Open and Distance Learning programs due
mainly to the spread of the unfortunate pandemic, COVID-19
(coronavirus), as they shall find both the electronic and print copies
indispensable.
Pedagogical features
Each chapter begins with an overview that aptly summarises the chapter
contents and a list of keywords similar to the chapter objectives. In every
chapter, there is at least an exemplary material (case study) with title, data,
research question and hypothesis that are embedded in the text. There are
screenshots as figures that explicitly illustrate procedural step by step
execution of the analysis covered in each chapter.
The general approach of the book to ensure that every user attains the
needful data analysis expertise with the most exceptional ease is captured
with the four fundamental features or structures of each chapter as listed.
i. Explanation of the Statistical Test – A statistical test is thoroughly
discussed in each chapter, clearly specifying the type of research
scenario that it is best used for analysing the observed data. At least,
one example of a suitable researchable topic is presented,
accompanied by a dataset.
ii. SPSS Dialog Boxes Selection Method – SPSS Dialog Boxes Point-
and-Click method is adopted with total screenshot illustrations to get
the data analysed. These complete illustrations make the book to be
second to none and a timeless first preference for tutors and students
of the courses that the book is meant, and for researchers, scientists
and individuals who deal with quantitative data analysis. How the
data in question are entered is meticulously specified for every
statistical test. On completion of each analysis, the output is
displayed without distortion.
iii. Interpretation of the Output – Exhaustive explanation of each
output is presented. Such detailed interpretation provides the user
with a much better understanding of the statistical test. It ensures that
the user gets every information that he would need for presenting,
22
interpreting and reporting his findings whenever he uses the
statistical test to analyse data from fieldwork.
iv. SPSS Syntax Methods – SPSS Statistics Syntax method is adopted
to analyse the same exemplary data. The requisite syntax is provided
with an illustrated simplified point-by-point guide to arrive at the
output. The SPSS syntax will enable users to become experts in
SPSS Statistics programming.
Final benefit
What if you could acquire the skills of data analysis with SPSS and
henceforth earn a meaningful living from it? Great. With this book, anyone
who wishes could independently acquire expert skills for data analysis,
using IBM SPSS Statistics; and gainfully earn a living from the new skills.
Take advantage of the rare golden opportunity straightaway.
23
Chapter 1
WELCOME TO IBM SPSS STATISTICS
Overview
The information communication and technology-led world bombards us
daily with enormous data on phenomena that demand correct analysis to
elicit optimal meaningfulness. In this chapter, IBM SPSS Statistics, the
most powerful, precise and valuable statistical software, is introduced to
ensure that each user practically engages in applying it with most
exceptional ease to analyse different sorts of data for specific purposes.
Data analytics is undergoing a revolutionary paradigm-shift from use of a
hand-held calculator to automated applications of statistical software.
Welcome to IBM SPSS Statistics introduces the new paradigm via the
process of actively learning the SPSS user interface by finger exercise that
the profession demands. Launching the software, contents and functions of
the SPSS Data Editor, Viewer window and Syntax Editor, as well as saving,
printing and retrieval of files are pragmatically demonstrated in a
pedagogically sound manner.
24
Equally presented herein are printing of Data files, Syntax files and of the
Output of statistical analysis. Learning of the Welcome to IBM SPSS
Statistics information will greatly facilitate easy and quick mastery of how
SPSS is practically applied in the analysis of each of the several statistical
tests in subsequent chapters of the book.
SPSS software was first launched in 1968. The software, SPSS, stands
for Statistical Package for the Social Sciences. In 2009, International
Business Machines Corporation (IBM) acquired SPSS and the software has
since then become popularly known as IBM SPSS Statistics officially.
Throughout this book, anywhere SPSS is seen, it refers to the IBM SPSS
Statistics. All types of data, irrespective of the source, can best be edited
and analysed with the SPSS Statistics. Data formats like plain text files (.txt
or .csv); relational databases; spreadsheets from Microsoft Excel or
OpenOffice; SAS; R, Python and Stata can all be opened and used easily
with SPSS.
The IBM SPSS Statistics is a landmark advancement in Information and
Communication Technology (ICT) that is radically changing the world
swiftly for better. IBM SPSS Statistics is indeed the most comprehensive
statistical software system for accurate analysis of all sorts of data, from the
least to the most complex in various fields (Kpolovie, 2018). With
application of SPSS, statistical analysis is most easily accessible for the
beginners and much more convenient for the experienced users in
comparison with some other statistical software. It is little wonder that
Kpolovie (2011; 2018) and Howitt and Cramer (2008) asserted that SPSS
statistics is by far the most widely employed computer package for data
analysis globally.
Learning how IBM SPSS Statistics is used is a highly transferable skill
that is usually a valuable asset in the labour market and in all spheres of
human endeavour. IBM SPSS Statistics is commonly used “at all levels
from students to specialist researchers, and in great many academic fields
and practical settings” Howitt and Cramer (2008, 4). SPSS statistics is an
incredible software for execution of advanced statistical analysis, text
analysis and a vast library of machine learning algorithms. It allows for
seamless integration with big data and deployment in practical applications.
The SPSS platform as presented in this book is extremely flexible, scalable
and most easy-to-use by everyone from all fields of life in the digital world
25
that we live. Acquiring the skills for use of IBM SPSS Statistics guarantees
the greatest possible analytical capability that is of inestimable value.
Unlike other statistical software that necessarily requires typing in
several lines of command syntax, SPSS allows for use of only simple menu
and dialog box selections to execute or perform analysis, even the most
complex ones. Entry of data and browsing are easily done in a simple and
efficient spreadsheet-like facility that IBM SPSS Statistics Data Editor
provides (Kpolovie 2018).
IBM (2019) summarised what the software does thus: “IBM SPSS
statistics is the world’s leading statistical software. It enables you to quickly
dig deeper into your data, making it a much more effective tool than
spreadsheets, databases or standard multi-dimensional tools for analysts.
SPSS Statistics excels at making sense of complex patterns and associations
– enabling users to draw conclusions and make predictions. And it’s fast –
handling tasks like data manipulation and statistical procedures in a third of
the time of many non-statistical programs.”
To purchase and install the IBM SPSS Statistics software, you are please
requested to visit IBM SPSS Statistics Websites at
https://www.ibm.com/products/spss-statistics. At the IBM SPSS statistics
website, it is possible for you to very easily subscribe to use the software
instantly. The subscription guarantees you immediate total access to the
most current IBM SPSS Statistics version at each time. The subscription
could be monthly, quarterly or yearly, depending on your choice. At the
IBM SPSS Website, you will also be able to start practicing IBM SPSS
statistics with the available Free Trial Version even before you either
subscribe or buy the software. For buying or subscription, special offer for
students is as well available at the Site. Once you purchase the IBM SPSS
Statistics software, it can automatically be installed online for your instant
use. A CD as backup of the particular version you bought will also be
shipped freely or at unimaginably small cost to you within a few days.
The entire analysis procedures and illustrations with screenshots in this
book are done with IBM SPSS Statistics Grad Pack 26.0 Premium which
is the most current and highest version of the software for now. Confidently,
there will not be any difficulty at all in applying the instructions on how to
analyse data with IBM SPSS Statistics as definitively presented in this
book, SPSS Statistics Excellent Guide, by those who may be using most
26
previous versions of the software as well as those who will be using
subsequent versions of IBM SPSS Statistics.
For each statistical analysis with IBM SPSS Statistics, there are ten
simple steps, one leading to the other as shown in Figure 1.0. The steps are:
switching the computer on, launching (opening) IBM SPSS, entering the
data, saving the data input, analysing the data, viewing the output, saving
the output, printing the output and data, exiting SPSS, and shutting down
the computer. These steps are illustrated explicitly with each statistical test
in this book.
27
All that is required to be done at each of the steps shall soon be
demonstrated for you to master. Apart from the fifth step, Analyze the
data, what the other steps require are about the same for different statistical
analysis execution. The fifth step varies because every statistical test has its
peculiar way of analysing the data, depending on the requisite menu or
dialog boxes selections that are most suitable for it. In all, the analysis
procedure will almost be like doing the same thing over and over for
various statistical approaches. Attainment of mastery knowledge in
personally applying IBM SPSS Statistics in data analysis is therefore
guaranteed. For each analysis, a thorough interpretation of the IBM SPSS
Statistics output is provided. There just could not be a better opportunity for
your becoming very skilful in data analysis with IBM SPSS Statistics than
as passionately presented in this book. Enjoy every bit of it.
28
29
The IBM SPSS Statistics, irrespective of the form, can be installed and
licensed in computers that run on different Operating Systems like
Windows, Mackintosh, Linux, HP-UX, Solaris and IBM AIX. Though the
presentations in this book are based on Windows OS, very little or no
additional efforts are needed to accurately and very easily apply them, using
computers that are based in other Operating Systems because IBM SPSS
Statistics is developed on Java programme.
30
LAUNCHING IBM SPSS STATISTICS
To start or launch IBM SPSS Statistics, simply switch on the computer that
has the software installed, allow it to completely boot up after which you
click the IBM SPSS Statistics 26 icon , click the IBM SPSS
Statistics icon, for whichever version of IBM SPSS Statistics that is
installed in your computer. If the computer does not yet have any version of
IBM SPSS Statistics installed, quickly visit the IBM SPSS Website at
https://www.ibm.com/products/spss-statistics to purchase it or subscribe
for it and get it installed in your computer, or use the Free Trial Version for
IBM SPSS Statistics that is available at the Website. A single click on the
IBM SPSS Statistics icon opens or starts IBM SPSS that displays the
IBM SPSS Statistics Version (Version 26.0 is used in this book) along
with its Copyright information as demonstrated in Figure 1.1. Shortly
afterwards, you will see Welcome to IBM SPSS Statistics as presented in
Figure 1.2.
31
Once the IBM SPSS Statistics software is opened, the very first page
that comes up and lasts for just a few minutes that enable you to very
carefully read it understandably, is the IMB SPSS Statistics Copyright and
registered trademarks information. The page alerts you that there is only
one company, the International Business Machines Corporation, which
owns the software and exclusively has the copyright of the software and
should therefore be properly referenced whenever SPSS Statistics is used. I
very strongly recommend that whenever you use SPSS Statistics software
for data analysis in an investigation, be it dissertation, thesis, supervised
research project, journal article, paper publication, professional research, or
book; you should accord IBM Corporation the glory of at least referencing
the Corporation that absolutely owns the copyright. Taking this book for
instance, the International Business Machines Corporation is duly
referenced under the Copyright section page that is strictly in line with the
express permission granted me by IBM Corporation to use the IBM SPSS®
Statistics screenshots and registered trademarks found throughout this book
(IBM SPSS Statistics Excellent Guide). The referencing of IBM could be
done as footnote at the very bottom of each page or at least of the first page
where any of IBM’s trademarks or copyrighted materials appeared and
again under the list of references of your work.
You may have read in the methodology and results sections of some
published research works that a particular version of SPSS was employed
for the data analysis, but no proper referencing or citation of the IBM SPSS
Statistics could be found on that page or in the list of references. Such use
of IBM Corporation’s materials without duly acknowledging and
appreciating the International Business Machines Corporation is not
scholarly. Ensure to do the needful at all time. In fact, the essence of the
IBM SPSS Statistics copyright appearing as the first page whenever you are
opening SPSS Statistics is to unconditionally remind you to always
acknowledge and properly reference IBM Corporation that owns the
copyright of the software whenever you use the SPSS Statistics to analyse
your data. At the first mention of IBM, SPSS or SPSS Statistics, you could
put an asterisk and explain it at the footnote that “IBM, SPSS or SPSS
Statistics is the property of IBM Corp. ® Copyright IBM Corporation and
its licensors 1989, 2017. IBM, IBM logo, ibm.com, and SPSS are
32
trademarks or registered trademarks of International Business Machines
Corp., registered in many jurisdictions worldwide” (IBM Corp., 2019).
As soon as the IBM SPSS Statistics copyright information in Figure 1.1
disappears, you must see the Welcome to IBM SPSS Statistics window or
dialog box that is exemplified in Figure 1.2.
At the top left of the start-up dialog widow (Welcome to IBM SPSS
Statistics), there is a rectangle containing two New Files that are named
New Dataset and New Database Query… Selecting New Dataset opens
an empty IBM SPSS Data Editor window for data entry. Double clicking
on the Database Query… welcomes you to the Database Wizard that can
help you to obtain data from a non-SPSS Statistics data source. In its dialog
33
box, you can choose the data source, specify which cases are retrieved,
aggregate and sort the data before retrieving. You can also specify the
variable names and properties in the Database Wizard. However, it is only
when the data source is connected to SPSS Statistics Server that all of the
features will be available. In the rectangle that is named Recent Files,
various data files that have been created in SPSS and saved in the computer
are listed for easy retrieval by clicking the Open another file. If you have
previously typed and saved some data files in SPSS and you need to work
on a particular file, just double click (clicking the file name twice in close
succession, using the mouse) or click the name of the file once and click
Open, or simply press Enter for the file to be opened. To open the
numerous default data files in IBM SPSS, just click the Sample Files
beside Recent Files. You can use the data saved in these Sample Files to
practice analysis of different statistical tests.
At the right of the Welcome to IBM SPSS Statistics window is a
rectangle that is named What’s New which allows for Smart Importing
and Exporting of data files from other sources into or out of IBM SPSS
Statistics. Below this box is another rectangle that allows you to see and use
a wide range of support resources by clicking Visit the Community.
The next box in Figure 1.2 is named Getting Started, which provides
two major functions of Get Help and Support and Get started with
tutorials. While clicking the first function enables you to get any type of
help that you want on how to apply SPSS, clicking the second (Get started
with tutorials) opens a special environment that provides you with tutorials
which practically show how data entry and analysis are done with IBM
SPSS Statistics for various statistical tests. You are encouraged to take the
special advantage that the Get started with tutorials offers, particularly if
this is your first time of using IBM SPSS Statistics.
Alternatively, you can simply proceed with your data entry and analysis
by clicking Close to close the Welcome to IBM SPSS Statistics window.
This action will lead you automatically to the IBM SPSS Statistics Data
Editor window. If you do not want the Welcome to IBM SPSS Statistics
dialog box or window to show at all when you open or launch IBM SPSS
Statistics subsequently, select the small square that is at the bottom right
corner of Figure 1.2 that is termed Don’t show this dialog in the future.
After that, close the Welcome to IBM SPSS Statistics window by either
34
clicking Close at the bottom right corner or the X mark at the extreme top
right corner of the window.
35
Figure 1.3 IBM SPSS Statistics Data Editor
36
2. Toolbar
3. Data View
4. Variable View.
Each of these components is very briefly described in this chapter.
MENU BAR
The row at the top of the IBM SPSS Statistics Data Editor, shown in
Figure 1.3, where different names of statistical functions are written is
known as the Menu Bar. Each of the named features in the Menu Bar is
used for execution of several functions as very briefly presented in Figures
1.4 to 1.14. A menu in IBM SPSS Statistics is a list from which the user
may select an operation to be performed, and the selection is often done
with the mouse or other pointing device under graphical user interface, or
may be controlled via the keyboard. To see what each menu in the Menu
Bar does, move your mouse pointer, known as cursor, to the menu and
click. Clicking on a menu displays all the menu items that are within it; and
in turn, a click on each menu item reveals or activates its sub-menu items or
opens its dialog box. A dialog box is a window in IBM SPSS Statistics
that allows the user to perform some simple actions like selecting certain
information in it as part of the procedures for execution of a given statistical
analysis. There are 11 menus in the Menu Bar of the IBM SPSS Statistics
Data Editor in Figure 1.3. The 11 menus in the Menu Bar are: File, Edit,
View, Data, Transform, Analyze, Graphs, Utilities, Extensions,
Window, and Help. The various menus or features in the Menu Bar of the
IBM SPSS Statistics Data Editor and the functions contained in each of
them are diagrammatically illustrated in Figures 1.4 to 1.14.
File
37
Figure 1.4 IBM SPSS Data Editor File
Edit
38
Figure 1.5 IBM SPSS Data Editor Edit
View
39
Figure 1.6 IBM SPSS Data Editor View
Data
40
Figure 1.7 IBM SPSS Data Editor Data
Transform
41
Figure 1.8 IBM SPSS Data Editor Transform
Analyze
42
Figure 1.9 IBM SPSS Data Editor Analyze
Graphs
43
Figure 1.10 IBM SPSS Data Editor Graphs
Utilities
Extensions
44
Figure 1.12 IBM SPSS Data Editor Extensions
Window
Help
TOOLBAR
45
Immediately under the Menu Bar in the IBM SPSS Data Editor shown in
Figure 1.3, is the Toolbar button that consists of 16 icons for easy
performance of different statistical operations. These icons serve as
shortcuts to facilities and features that are very often utilized in the menus.
Each of the 16 tools in the Toolbar is represented with an icon. The names
of the icons and their specific functions are very briefly given as follows:
1. Open data document: is used for opening previously saved
data files (if you are in Data Editor) or output files (if you are in the
output Viewer).
2. Save this document: is employed for saving of the current file
that you are working with. Where the file has not been saved already
before at the point of creation, it opens the Save Data As dialog box,
prompting you to give it a name and click Save or press the Enter
key on the Keyboard to get it saved.
3. Print: is adopted for printing whichever file that you are
currently working on, be it data file or output file up to the desired
number of copies that you will instruct your printer to print. For an
output file, it will print the entire file unless you have selected
specific output tables alone for printing.
4. Recall recently used dialog: produces a list of 12 dialog boxes;
and activates or opens any particular one that you select. It enables
you to execute any analysis that you have done before, yet another
time, perhaps for correction of a suspected mistake; or for the crucial
sake of repeating the analysis to acquire mastery knowledge. Usually,
I tell my students that repetition is the key to mastery learning of
IBM SPSS Statistics.
46
7. Go to case: a click on this icon produces a dialog box in which
you quickly type the case number (a serial number) and click Go of a
particular case or respondent that you want to see and confirm his
scores in a large data set. This saves you the time and the
inconvenience of doing a long scrolling down or up in search of the
respondent.
47
11. Find: enables the user to search for a given value or word in
the Data Edit or Viewer or window whenever the need arises to look
for it and perhaps correct it if it was entered wrongly or if its relative
location in data set was needed.
12. Split File: opens a dialog box for splitting data on a variable
into desired categories for separate or comparative analysis.
13. Select Case: where only some cases are needed to perform
specific analysis instead of using the entire cases, the Select Case
icon is clicked to produce a dialog box for the purpose. It allows for
selection of the specific cases out of the large data set for the
analysis.
15. Use Variable Set: clicking this icon allows for selection of sets
of variables for analysis such that two or more sets of variables can
be compared or regressed at once.
48
DATA VIEW
At the extreme bottom left corner of the IBM SPSS Statistics Data Editor,
are two possible views that are named Data View and Variable View,
which must be used respectively for entering the data and specifying
characteristics of the variables before the data analysis proper. The Data
View is a spreadsheet-style interface made up of numerous cells that are
formed with columns and rows for data entry. Each of the columns is by
default named Var (short form for Variable) because each column is meant
for entry of numerical information or data about a given variable. The rows
are by default serially numbered 1, 2, 3, 4, 5, and so on, to represent the
subjects, participants or respondents in a particular investigation and from
whom the data were observed. The numerous cells created by the columns
and rows are where data on the independent and dependent variables under
investigation are entered or typed-in. The very first major thing to do in
order to analyse data with IBM SPSS Statistics is the entering of the data
which is done in the Data View. In fact, the IBM SPSS Statistics Data
Editor that you see once you launch or start SPSS Statistics and close the
superimposed dialog box is the Data View because the various versions of
SPSS for Windows have the Data View set as the default for inputting of
data.
Suppose that an investigator interested in finding out whether gender
difference exists in primary school students’ Basic Science Concepts,
administered a standardized test of 60 items on Basic Science Concepts to a
set of randomly sampled K-6 students, and they obtained the scores listed in
Table 1.2. The sample size that the investigator studied is 69 as will later be
seen in the first table of Chapter 3. Only scores of the first 20 participants
are used here in this Table 1.2 for illustration of the Data View.
49
5 Male 40
6 Male 50
7 Female 10
8 Male 40
9 Male 20
10 Female 20
11 Female 30
12 Female 50
13 Male 40
14 Male 30
15 Male 10
16 Male 30
17 Female 50
18 Female 40
19 Female 40
20 Male 20
The very first step in analysing these data is entering the data in the IBM
SPSS Data View of the Data Editor. It must very quickly be noted here
that in data entry, the categorical or grouping variable, Gender in this case,
must be coded such that the males are assigned 1 each and the females are
assigned 2 each (or having the males coded 2 and females coded 1
consistently will not make any difference). Note that in this example, males
have code 1 and females have the code 2. This has to be done to ensure that
only numerical data are entered in the Data View.
DATA ENTRY
50
Note emphatically that data entry is done in the Data View of the IBM
SPSS Statistics Data Editor before specifying the variables in Variable
View that will be treated in the next section shortly afterwards. This is
partly because IBM SPSS Statistics has the Data View as the default for
data entry and partly because it is the Data View that is under consideration
at the moment. The entire data on the independent variable which is
Gender with males coded 1 and females coded 2 will be entered under the
first column. Once a value is entered in this column, SPSS will
automatically describe or name the column as Var00001. After very
carefully entering the data on Gender (the categorical variable with males
as 1 and females as 2), the Scores that constitute data on the dependent
variable that is Basic Science Concepts will be entered under the second
column in such a careful manner that the two values (gender and score) for
each participant form one row. Once the first value or score is entered under
the second column, SPSS automatically labels the column as Var00002.
Take note of the Var00001 and Var00002 as they will later change to
Gender and Scores, respectively, after naming them in Variable View.
To perform the actual data entry, ensure that the cursor (mouse pointer)
is in the topmost cell of the first column. Then, type the value that indicates
each participant’s gender and press the Enter button. Do not fail to press
Enter after typing each value. You must have observed that as you type a
value and press Enter, the cursor moves down to the next immediate row. If
you have not done the typing and pressing of the Enter key yet; do it now.
Type 1 and press Enter, 2 and press Enter, 2 and press Enter, 2 and
press Enter, 1 and press Enter, 1 and press Enter until all the values on
Gender are entered without a mistake. Next, move the cursor to the
topmost cell under the second column and click or move the cursor to that
cell by pressing the arrow right ( ) button once and pressing the arrow up (
) button and holding it down until the cursor gets to and remains in the
topmost cell of the second column. Very carefully enter the scores from that
of the first participant as in Table 1.2. That is, type 10 and press Enter, 20
and press Enter, 30 and press Enter, 10 and press Enter, 40 and press
Enter, and so on till the last score, 20 and press Enter. What you have on
your screen will be as in Figure 1.15.
51
Figure 1.15 Data View when the variables are not yet specified
VARIABLE VIEW
The Variable View is a very special window in SPSS Statistics Data
Editor that is used for creation and specification of the variables under
study. It is the Variable View that allows for description of the
characteristics of the variables on which data are collected, entered and are
to be analysed. The description includes provision of name, label, values,
measure and role as well as for the specification of the type, width, missing
value, columns and alignment for each of the variables. To access the
Variable View, place the cursor on and click the Variable View that is
beside the Data View at the extreme bottom left of the Data Editor. This
simple action opens the Variable View as shown in Figure 1.16.
52
Figure 1.16 Variable View of the IBM SPSS Statistics Data Editor
Recall that when data were entered in the Data View, SPSS
automatically provided default names for the two variables. The first
column that is the independent variable, where the categorical data, were
entered was named VAR00001 and the second column where scores on the
dependent variable, the students’ Basic Science Concepts, were entered was
named VAR00002. These VAR00001 and VAR00002 are the temporary
names for the two variables under study that you are seeing in the Variable
View under Name. The essence of opening the Variable View is to
describe and specify the variables by giving them their actual names. So,
click on VAR00001 in the first row and type the real name that best
describes that variable, which is Gender. Next, accord a similar treatment
to the VAR00002 in the second row by clicking it and typing the actual
name for this variable, which is Scores.
The second column in the Variable View is called Type and has rightly
been described as the Numeric data for the two variables (Gender and
Scores). While the third column, Width, indicates the number of digits or
53
characters that are displayed for each of the variables to be 8, the fourth
column, Decimals, has shown that both of the variables will have 2 decimal
places. The fifth column, Label, requires you to assign the most suitable
label that best describes each of the variables. The first variable which is
Gender can best be labelled Gender, but just for the purpose of illustration,
label it as Sex; and label the second variable, Scores, as Basic science
concept scores. These labels will appear in the output of the analysis. You
may wish to know why apart from providing names to the variables, SPSS
Variable View still wants you to provide label for each variable. The name
of a variable cannot begin with a number or numeral and the variable name
cannot accept any space; but the label can commence with a numeral and it
can accept space.
The sixth column, Values, is where you provide crucial additional
information to each categorical variable by stating the value you coded each
category with and defining each category by providing its specific label. To
illustrate this with the categorical variable, Gender, move the cursor to the
first row under the column, Values, that is written None and click. This
makes a small blue box with three dots which is an ellipsis button to
appear beside the None as shown in Figure 1.16.
Then, click on the small blue box (i.e., the ellipsis button) to produce the
dialog box that requires you to type in the Value and Label for each
category of the Gender as in Figure 1.17. Recall that while males were
coded with the value 1, females were coded with the value 2.
54
Figure 1.17 Values dialog box
Then, type 1 in the box beside Value and Male in the box beside Label
as exhibited in Figure 1.18.
Figure 1.18 Values dialog box with value and label for male.
Next, click Add for the entered male’s value and label to be added into
the big box beside Add as shown in Figure 1.19.
55
Figure 1.19 Value Labels dialog box after clicking Add for male
To specify the coded value and label for the female’s category of
Gender, type 2 in the box beside Value and Female in the box beside
Label. When done, the Value Labels dialog box will look like what is in
Figure 1.20.
Figure 1.20 Values dialog box with value and label for female
56
The value and label that have been typed cannot enter into the analysis
unless you click Add. So, click Add to move the entered value and label for
females from where they are into the big box beside Add. When done, the
Value Labels window will become like what is exhibited in Figure 1.21.
Figure 1.21 Value Labels dialog box after clicking Add for female
After clicking OK, you are done with the Values; and the None that was
in the cell for Gender row against Values column gets replaced with the
57
name for one of the categories, in this case, Male.
The seventh column in the Variable View window is known as Missing
which is used to indicate the number and percentage of missing values or
cases from the data. Any value of the variable that is not available is treated
as a missing value. For the current example data that were entered in the
Data View, there is no value that was missing, so the Missing column has
rightly recorded None for Gender and None for Scores. The eighth column
of Variable View is called Columns that is used to show the width of each
displayed column at the point of data entry. By default, the Columns has
the width of 8 characters, and is so stated for each of the two variables.
Align is the ninth column of the Variable View; and it is used to indicate
how data entered for each variable will be aligned either to the left, right or
centre of the cells in the Data View. The tenth column of Variable View is
termed Measure. This specifies the scale of measurement at which data
were collect for each variable. There are three possible options that data on
a variable could take. They are Nominal, Ordinal, and Scale. The Scale is
used to represent data collected at interval scale of measurement or at ratio
scale of measurement. For the example under consideration, data on Scores
that the K-6 students got on Basic Science Concepts are collected at the
interval level of measurement that SPSS describes in the Variable View as
Scale.
Take the cursor to the cell for Gender against Measure that is written
Unknown, click it and you will see a small dropdown dialog box in which
you select Nominal. Then, put the cursor on Unknown at the cell for
Scores against Measure, click it to see its dropdown dialog box and select
Scale as shown in Figure 1.23. The last column of the Variable View is
Role for pre-specification of variables as either Input, Target, Both, None,
Partition or Split (shown in Figure 1.23) for different uses in certain dialog
boxes.
58
Figure 1.23 Variable View window with Measure specified and options for Role shown
Haven gone through the various fields in the Variable View, it is time to
again switch over to the Data View in the IBM SPSS Statistics Data
Editor. On getting to the Data View, you will notice very easily and
interestingly that the data which when you entered in Figure 1.15 had
Var00001 and Var00002 as names for the two variables in question, are
now having the actual names of the variables as Gender and Scores
respectively in the first and second columns of the Data View. Gender and
Scores have automatically replaced VAR00001 and VAR00002 because of
the specification of the characteristics of the variables that was done in the
Variable View. It shows the robustness and complementary nature of the
Data View and Variable View of the IBM SPSS Statistics Data Editor.
To switch over from Variable View to Data View, simply move the cursor
to and click Data View that is at the bottommost and extreme left corner of
the Variable View. When done, the Data View now looks like what is in
Figure 1.24.
59
Figure 1.24 Data View after specifying the variables
60
the cursor to Gender in the Data View and click. This click highlights the
entire column for Gender as shown in Figure 1.25.
61
With the Gender column highlighted, move cursor to the Toolbar and
click the Value Labels icon, . With this click, the actual labels (Male
and Female) replaces the coded values (1 and 2), respectively as exhibited
in Figure 1.26. With the Gender column highlighted, and you again click
the Value Labels tool in the Toolbar, the labels Male and Female gest
replaced with the coded values of 1 and 2, respectively, as shown earlier in
Figure 1.25. This is a further demonstration of the flexibility, robustness
and complementary nature of the Data View and Variable View in IBM
SPSS Statistics Data Editor.
62
Figure 1.26 Data View with the value labels for Gender
63
When I began description of the IBM SPSS Statistics Data Editor, I
mentioned that there are three major windows that IBM SPSS Statistics
uses, namely:
IBM SPSS Statistics Data Editor
IBM SPSS Statistics Viewer
IBM SPSS Statistics Syntax Editor.
The IBM SPSS Statistics Viewer is a most special window for the
viewing of all results, preferably termed Output of data analysis done with
the software (IBM SPSS Statistics). Very aptly, the SPSS Statistics Viewer
window can be referred to as the SPSS Output window. For the IBM SPSS
Statistics Viewer to be used, data must have first been analysed. So, let me
use very simple descriptive analysis of the data in Table 1.2 that have been
entered in Figure 1.24 to very quickly illustrate the IBM SPSS Viewer
window by running a very quick analysis.
To execute very quick descriptive analysis of the data in Figure 1.24,
there are just two steps for each of the variables:
For Gender, the first step is to move the cursor to Gender and click
to highlight the Gender column as can be seen in Figure 1.27.
64
Figure 1.27 Data View with Gender and Scores highlighted
The second step is to move the cursor to and click the Run
65
descriptive statistics analysis for the data and produces the Output or
results that are demonstrated in the first part of Output 1.1.
Then, get back to the IBM SPSS Statistics Data Editor that is containing
the dataset as in Figure 1.24. Repeat the two steps action, but for Scores
this time. That is, while on the SPSS Data Editor containing the dataset,
move cursor to and click Scores to highlight the Scores column as shown in
Figure 1.27. This is the first step. Then, for the second step, move cursor to
and click the Run descriptive statistics tool in the Toolbar while the
Scores column is highlighted. With this click, quick descriptive statistics
analysis is executed for the Basic Science Concepts data, named Scores;
and the results produced in the IMB SPSS Statistics Viewer as displayed
in the second part of Output 1.1.
Also, the quick Descriptive Statistics can be done at once for both
variables instead of doing it for each of the variables. Click to select both
Gender and Scores for the two columns to be highlighted as shown in
Figure 1.27. Next, move cursor to and click , the Run descriptive
statistics tool in the Toolbar. With this single click, IBM SPSS Statistics
instantly completes the analysis and produces the results in the IBM SPSS
Statistics Viewer as presented in Output 1.1. The window where outputs
are displayed in IBM SPSS Statistics is the IBM SPSS Statistics Viewer.
66
67
INTERPRETATION OF OUTPUT 1.1 AS
ILLUSTRATION OF SPSS VIEWER
IBM SPSS Statistics always presents the results, better known and more
frequently called Output, in the Viewer window. Sections of the output are
summarized in user-friendly tables. The first table within the output
indicates Frequencies Statistics for the variable Sex (recall when in the
Label column of the Variable View, Sex was typed in as the Label instead
of Gender). It shows the valid and invalid number of cases (N) for sex. The
Mode, Range, Minimum and Maximum are also shown for Sex. The second
table for Sex within the first part of the output has indicated the Frequency,
Percent, Valid Percent and Cumulative Percent for all the cases that are
male and female.
The third table, which is the second part of the Output 1.1, has given the
Frequencies Statistics for Basic Science Concepts, again recall when in
the Label column of Variable View; the Basic Science Concepts was
typed in as the Label instead of Scores. The Valid (20) and Missing (0)
number of cases (N) are shown in this table. Also, the Mean (29.5000),
Median (30.0000), Standard Deviation (13.94538), Range (40), Minimum
(10), and Maximum (50) are given as the total (male and female students
68
together) descriptive statistics for the variable, Basic Science Concepts
(i.e., the Scores).
Haven looked at the output of the data analysis in the IBM SPSS
Statistics Viewer, it will be optimally easy for best understanding of the
importance of the IBM SPSS Statistics Viewer window now. The IBM
SPSS Statistics Viewer, like the other two major windows in SPSS, has
Menu Bar and Toolbar. The Menu Bar has 13 distinct functions or menu
items that are listed as File, Edit, View, Data, Transform, Insert, Format,
Analyze, Graphs, Utilities, Extensions, Window, and Help. To see the
several statistical mechanisms that each of the menus in the Menu Bar is
used for, move the cursor to it and click.
The Toolbar in IBM SPSS Statistics Viewer window has 16 tools that
are represented with icons to provide very quick access to commonly used
functions in the Menu Bar. They are:
3. Print
5. Export
9. Go to Data
10. Go to Case
11. Go to Variable
69
12. Variables
For a brief understanding of what each of these Toolbar tools does in the
SPSS Viewer, move your cursor to it and click to have its dropdown dialog
box that outlines the statistical operations that can be performed with the
tool. Most of the functions of the Menu Bar items and Toolbar tools in IBM
SPSS Viewer window are much like what the Menu Bar menus and
Toolbar tools do in the IBM SPSS Statistics Data View that have been
described earlier.
In addition to the Menu Bar and Toolbar, it is most glaring that the IBM
SPSS Statistics Viewer window has two sides, the left part and the right
part. The left side provides an outline that represents the contents of the
statistical analysis done in a tree-like structure that produced the output.
The right side presents the actual results or output, summarized in most
suitable research-report-ready tables, charts, graphs, models and texts.
Subsequently, it is the information presented in the right-hand section of the
IBM SPSS Viewer window that will be presented and explained briefly for
each statistical test in this book. Just know that the results in the right part
are in accordance with the decision-tree-like structure in the left part of the
IBM SPSS Statistics Viewer for any specific analysis done.
70
IBM SPSS Data Editor
IBM SPSS Viewer
IBM SPSS Syntax Editor.
IBM SPSS Statistics Syntax Editor is a most powerful window that
executes statistical analysis by simply typing in the SPSS Programme
Commands, known as SPSS Statistics Syntax for running each statistical
test accurately with greatest ease and swiftness. All that it typically requires
is your mastery of the SPSS Programme Commands necessary for
performing any given statistical analysis. You do not need to worry yourself
any bit on how to know those Commands because all the necessary SPSS
Syntax or SPSS statistical command jargons are provided for you in this
book. IBM SPSS Statistics has so much simplified the statistical
programme Commands, as presented in this book, that you only need to get
a copy of this text and meticulously enter the Syntax as explicitly provided
herein. The only modification that you will be doing is replacing the names
of the variables used for illustration in this book with names of the actual
variables that you collected data on in your investigation.
Most likely, there cannot be a better way of learning how to correctly
write IBM SPSS Statistics Programming Language (known as
Command Syntax) than as warmly articulated with precision in this book.
For each statistical test, once you start entering the IBM SPSS Command
Syntax as in this book, the SPSS Statistics software will also prompt you to
speedily type the Commands into the Syntax Editor by changing the text
colour and providing little drop-down menus from time-to-time for you to
choose the word(s), term(s) or phrase(s) that you intend to type. You are
bound to most speedily learn and master how to write IBM SPSS
Statistical Syntax as illustrated with exciting amazement. You are indeed
welcome to IBM SPSS Statistics, the only statistical home that instantly
catches and sustains your total attention until you most successfully
conclude analysing all sorts of research data.
Therefore, data analysis with IMB SPSS Statistics can conveniently and
most comfortably be done in two ways.
1. Data Editor Method: Statistical analysis is done in this method with
SPSS dialog boxes selection and clicking via the Data Editor
window. This method of data analysis is also referred to as SPSS
Dialog Boxes Point and Click.
71
2. Syntax Editor Method: Data analysis is done in this method by
typing in a few SPSS Statistics Programming Language, known as
Command Syntax or simply as Syntax, via activation of the Syntax
Editor window.
The two methods of data analysis with the use of IBM SPSS Statistics
share two main common features or characteristics. Firstly, the data to be
analysed must be entered and well labelled as already explained under Data
View and Variable View of the Data Editor. Secondly, the output of data
analysis done in both of the methods must always be shown in the IBM
SPSS Statistics Viewer window as discussed already.
Reason for why the Data Editor approach and the Syntax Editor
approach have much in common is because in both methods, the analysis is
actually executed on the basis of the same principle of IBM SPSS
Statistical Programming Language. The only difference is that while in
the Data Editor technique, the Command Syntax is written behind the
correct dialog boxes for selection; in the Syntax Editor technique, the user
of IBM SPSS has to key in the programming Command Syntax that IBM
SPSS Statistics has most carefully written and made available for the
public (IBM Corp., 2019) as aptly demonstrated in the current book.
72
Figure 1.28: Data Editor indicating File and New dropdown menus
73
Alternatively, move cursor to the Menu Bar and click Window; and in
the pulldown dialog box of the Window, click Go to Designated Syntax
Window. The actions of clicking Window and clicking Go to Designated
Syntax Window immediately opens the IBM SPSS Statistics Syntax
Editor that is shown in Figure 1.29.
The fifth step demands that you carefully key in the IBM SPSS
Statistics Programme Commands (i.e., Syntax) for the specific
statistical analysis which you are subjecting the entered data to. It is
in the IBM SPSS Statistics Syntax Editor shown in Figure 1.29 that
you type in or enter the Syntax meant for the statistical test that you
want to do.
To execute the simple descriptive statistics for the data on K-6 students’
Basic Science Concepts by Gender in Table 1.2 and Figure 1.24, type the
following Syntax or Commands presented in Syntax 1.1 into the Syntax
Editor.
74
After entering the requisite Syntax in the IBM SPSS Statistics Syntax
Editor, it will look like Figure 1.30.
Figure 1.30 IBM SPSS Statistics Syntax Editor with the entered Syntax
The sixth step is the final step and can be done in either of two
possible ways. First, select the Syntax that you have entered into the
SPSS Syntax Editor, move the cursor to the Toolbar and click ,
which is the tool or icon for Run Selection. The moment this is done,
IBM SPSS Statistics completely performs the analysis and produces
75
the output in the IBM SPSS Statistics Viewer on your screen in split
seconds as shown in Output 1.2.
The second way of performing the sixth step is by moving the cursor to
the Menu Bar and clicking Run. With this click, the Run dropdown dialog
box as in Figure 1.31 opens. In the Run dropdown dialog box, click All. By
clicking All, IBM SPSS Statistics totally executes the analysis and presents
the results in the IBM SPSS Statistics Viewer instantly on your screen as
displayed in Output 1.2.
76
The results of the descriptive statistics on students’ Basic Science
Concepts in Output 1.2 when the data were analysed with IBM SPSS
77
Statistics Syntax Editor method and in Output 1.1 that the same data were
analysed with IBM SPSS Statistics Data Editor method are exactly the
same. So, it is left for you to use any of the two methods of analysing data
with IBM SPSS Statistics. Let me reiterate that only the right side of the
SPSS Viewer is presented in the Output 1.2 as that point has earlier been
made; and that all subsequent analysis will have only the results in the right
part of the IBM SPSS Statistics Viewer displayed. Irrespective of the
method that you would choose to use, it is better to study and master the
two methods (SPSS Data Editor method and SPSS Syntax Editor method)
of data analysis with IBM SPSS Statistics as excellently treated definitively
in this book.
78
vi. Press the Enter key, then the Spacebar key, preferably the Tab key
on the Keyboard whenever a Command Syntax is getting too long to
be contained in a single line.
SAVING OF FILES
In IBM SPSS Statistics, there are three types of data, each of which
deserves to be saved as a file in a different environment with unique dialog
boxes and distinct extension to allow for easy retrieval and reference. If a
Data, Syntax or Output file that you used for a statistical analysis is not
saved, it cannot be readily retrieved at some other time or a later date for
use when the need arises. The three types of files in accordance with the
79
three major windows in IBM SPSS Statistics that should typically be saved
uniquely are:
Data file
Syntax file
Output file.
While the Data file can solely be saved when working in IBM SPSS
Statistics Data Editor window, the Syntax file can be saved exclusively
when working in IBM SPSS Statistics Syntax Editor window, and the
Output file can be saved outrightly only when working in IBM SPSS
Statistics Viewer window. The file extension for saving a Data file is .sav.
The file extension for saving a Syntax file is .sps. The file extension for
saving an Output file is .spv. You do not mandatorily have to type the
extension when naming a file for saving, because even when you do not
type the file extension, IBM SPSS Statistics will automatically add the
extension to complete the file name to enable the saving. Saving of file in
each of the three windows is only possible when the specific window and a
particular file are activated (i.e., when you are currently working on it or
when it is actively open). You can only save a Data file when you are
working on it in either Data View or Variable View in the IBM SPSS
Statistics Data Editor. Similarly, a Syntax file can only be saved while
you are working on it in the IBM SPSS Statistics Syntax Editor. Finally, it
is the moment that you are on an Output in the IBM SPSS Statistics
Viewer window that you can save the Output file.
80
Figure 1.32 Data Editor showing File in the Menu Bar
Step 2: Place the cursor on and click File in the Data Editor to display the
File dropdown dialog box as in Figure 1.33.
81
Figure 1.33 File drop-down dialog box in Data Editor
Step 3: Place cursor on and click Save As… in the File dropdown dialog
box for the Save Data As dialog box to appear as shown in Figure 1.34
82
Figure 1.34 Save Data As dialog box for Chapter 1 Table 2 Data.sav
Step 4: Type a suitable name for the File in the panel at the right of File
name. For instance, type Chapter 1 Table 2 Data.sav as the name for the
students’ Basic Science Concepts by Gender data.
Step 5: Either press Enter on the Key Board or move cursor to and click
Save that is located at the right close to the File name that you have just
entered. With this, the Chapter 1 Table 2 Data.sav has been created and
saved securely for future use.
Step 6: If you do more work on the file such as editing it or adding more
data after the file has been created and saved, move the cursor to and click
the Save icon on the Toolbar or press Ctrl and S simultaneously on the
Keyboard. With this, the file has been saved with the modification.
83
Figure 1.24. That is, in the SPSS Data Editor containing the data in
question, you clicked File, selected New, and click Syntax, after which the
Syntax 1.1 was entered into the SPSS Syntax Editor.
Make sure that you are in the Syntax Editor with the right Syntax or
statistical programme Commands entered as shown in Figure 1.30.
Alternatively, let me quickly show you another very easy way of
activating Syntax Editor.
From the Data Editor that you have entered the data to be analysed,
move cursor to the Menu Bar and click Window. This click makes
the Window dropdown dialog box to open as exemplified in Figure
1.34.
84
Carefully enter the SPSS programme Commands or Syntax used to
demonstrate quick Descriptive Statistics for data in Table 1.1 or
Figure 1.24. On completion of entering the SPSS statistical
Commands, the Syntax Editor will look as in Figure 1.36.
Figure 1.36 Syntax Editor with quick Descriptive Statistics Commands entered
85
Figure 1.37 Save Syntax As dialog box for Chapter 1 Syntax 1.sps
Enter a most suitable name for the Syntax file in the field at the right
of File name which is for that purpose. In this example, the
appropriate name to enter is Chapter 1 Syntax 1.sps as shown in
Figure 1.38.
86
Figure 1.38 Save Syntax As with file name entered
Finally, move cursor to, and click Save in the dialog box or press the
Enter key on the Keyboard. This saves the Syntax file as Chapter 1
Syntax 1.sps securely for easy access in the future whenever the
need arises.
87
programming language with which Data file and Output file are
respectively saved in the Data Editor and Viewer window. It is because of
the difference in the programming language used uniquely by each of the
three major IBM SPSS Statistics windows that a Data file must always
end with the extension .sav; a Syntax file must end always with the
extension .sps; and an Output file must always end with the extension .spv.
Follow the outlined steps to save the Output.
Step 1: Ensure that you are in the Viewer window, viewing the results of the
analysed data as earlier illustrated in Output 1.1 or in Output 1.2.
Step 2: Move cursor to the Toolbar and click , the icon or tool for Save
this document and the Save Output As dialog box will appear as in Figure
1.39.
Note that you can also get to the Save Output As dialog box that is
presented in Figure 1.39 via File in the IBM SPSS Statistics Viewer
window. To do this, while viewing your Output in the Viewer window,
88
move cursor to the Menu Bar and click File which is the first menu in the
Menu Bar and located at the topmost left extreme corner, to open File
dropdown dialog box as shown in Figure 1.40.
Figure 1.40 File dropdown dialog box in IBM SPSS Statistics Viewer
Then, move cursor to and click Save As… to have the Save Output As
dialog box as exhibited in Figure 1.39.
Step 3: In the File name panel that is like a row just at the right of File
name, type the name with which you want the Output to be saved. In this
example, the preferred name is Chapter 1 Output 1. So, type in Chapter 1
Output 1.spv as shown in Figure 1.41.
89
Figure 1.41 Save Output As dialog box with the file name
Step 4: Finally, press Enter on the Keyboard or move cursor to and click
Save that is just at the right of the File name panel where you typed in the
name of the output file within the Save Output As dialog box. With this,
the Output file has been securely saved for future retrieval and use.
PRINTING OF FILES
The printing of a file in IBM SPSS Statistics demands that you first
activate the very file that you want to print. That is, to print an Output file
like the one that is currently active (Chapter 1 Output 1.spv) in the IBM
SPSS Statistics Viewer on your screen, you must be in the Viewer
window. To print a Data file, you must activate the Data Editor and ensure
that the particular dataset you want to print is the one showing on the
screen. Similarly, Syntax file printing demands activation of the Syntax
Editor and opening of the specific Syntax file that you are interested in
printing. The activation of the particular window from which you want to
print a file is necessitated by the earlier emphasized point that each of the
90
three major windows (Data Editor, Viewer, and Syntax Editor) in IBM
SPSS Statistics operates on its unique programming language.
and click the tool or icon for Print for the Print dialog box to open as
shown Figure 1.42.
The Print dialog box for Output shows the name of the printer that is
currently set as the default printer for the computer. One of the two possible
Print Range, and is highlighted to let you
know whether what will be printed is All visible output or Selected
output. If no table in the Output was selected before your clicking the
Print icon , then the default of what will be printed is All visible
output. But if at least one table in the Output was selected when you
clicked the Print icon, then the Selected output will be the default possible
91
printing range that the printer will print for you. Also, in the Print dialog
box is the option for the number of copies that you want to get printed thus
. One copy is the default, but if you want more copies,
say 5 to be printed, you move the cursor to and click the up-pointed arrow
to increase the Number of copies till it gets to 5 thus
. There is as well a tick in the very small box beside Collate indicating that
the printer will while printing, collate (arrange) the papers on which the
printing will be done according to page-numbers for the 5 sets. The final
step in the printing is just for you to move cursor to and click OK in the
Print dialog box. With this, the printer prints out the Output.
92
Figure 1.43 File drop-down menu showing Print
In the File dropdown dialog box, move cursor to and click Print… for
the Print… dialog box to open, and you perform the actions spelt out in the
previous paragraph to get your Output printed.
93
starting from the IBM SPSS Statistics Viewer window where the Output
in question is located, the printing of a Data file demands that you start
from the Data View of the IBM SPSS Statistics Data Editor where the
Data file that you want to print is.
Be sure that you are right now at the K-6 students’ data saved as
Chapter 1 Table 2 Data.sav in the Data View of IBM SPSS Statistics
Data Editor. Click in the Toolbar to have the Print dialog box shown
in Figure 1.44. Then, click OK to have one copy of the Data file printed. If
it is five copies that you want to print for instance, increase the Number of
copies by clicking the up-pointed arrow beside it till it gets to 5; then click
OK in the dialog box or press Enter key on the Keyboard.
Syntax Editor, click the Print icon in the Toolbar to get the Print
dialog box and press Enter key on the Keyboard or click OK button in the
dialog box. With this, a copy of the Syntax file gets printed. You can repeat
94
the process to get any number of copies of the Syntax file that you want,
printed.
RETRIEVAL OF FILES
The entire essence of saving files in IBM SPSS Statistics is to guarantee
automatic retrieval as at when needed. Retrieval of a saved file can only be
done from the specific window where it was saved. That is, while retrieval
of a Data file should better be done from the IBM SPSS Statistics Data
Editor, retrieval of a Syntax file is best done from the SPSS Syntax
Editor, and retrieval of an Output file should necessarily be done from the
SPSS Viewer window.
Step 1. Move cursor to the Menu Bar in the IBM SPSS Statistics Data
Editor and click File to have its dropdown dialog box in which you move
the cursor to Open for the Open dropdown dialog box to appear as
illustrated in Figure 1.45.
95
Figure 1.45 File and Open drop-down dialog boxes
Step 2. Click Data in the Open dropdown dialog box. This produces the
Open Data dialog box presented as demonstrated in Figure 1.46.
96
Figure 1.46 Open Data dialog box for Chapter 1 Table 2 Data.sav
It is not only through clicking File in the SPSS Data Editor that you
can reach the Open Data dialog box. Alternatively, move cursor to Toolbar
and click the Open data document icon or tool for the Open Data
dialog box to be displayed as in Figure 1.46.
Step 3. Look for the name of the file you want to open, place the cursor on
it and double-click (press the left button on the mouse in very close
succession) or click the file name once and click Open or press the Enter
key on the Keyboard. In this case, the file name is Chapter 1 Table 2
Data.sav. Double-clicking on the file name or clicking the file name once
and either pressing Enter or clicking Open instantly retrieves the file in
question. Note that it is not possible to find a Syntax file or an Output file
within the Open Data dialog box. In fact, the title of the dialog box alone
(‘Open Data dialog box’) clearly stipulates that it is a Data file previously
created and saved that can and should be found in and retrieved via the
Open Data dialog box.
97
Retrieving Syntax File
A few steps similar to those for retrieving Data files are required for the
retrieval of saved IBM SPSS Statistics Syntax files from the SPSS Syntax
Editor for use at any moment.
Step 2. Select Open for the Open dropdown dialog box to appear, showing
Syntax as one of the options that could be selected in it. This is presented in
Figure 1.47.
Figure 1.47 File and Open dropdown dialog box showing Syntax
Step 3. Click Syntax… This action makes the Open Syntax dialog box to
appear as in Figure 1.48.
98
Figure 1.48 Open Syntax dialog box
Figure 1.49 Window dropdown dialog box showing Go to Designated Syntax Window
99
Figure 1.50 IBM SPSS Statistics Syntax Editor
Move cursor to the Toolbar in the SPSS Syntax Editor and click the
icon for Open Syntax to have its dialog box as earlier shown
in Figure 1.48.
Step 4. Place the cursor on and double-click the file name for the particular
Syntax that you want to retrieve (Chapter 1 Syntax 1.sps in the current
example). Apart from double-clicking on the file name, just click it once
and press the Enter key on the Keyboard or click Open in the Open
Syntax dialog box to instantly open the Chapter 1 Syntax 1.sps file.
Step 1. Ensure that you are in the IBM SPSS Statistics Viewer window to
be able to retrieve an Output file duly saved. You may want to know how
to get to the Viewer window. Very simple, click Window in the Menu Bar
to get its dropdown dialog box as in Figure 1.51.
100
Figure 1.51 Window dropdown dialog box showing Go to Designated Viewer Window
Step 3. Move cursor to and click the Open output document icon in
the Toolbar of the SPSS Viewer window. This click opens the Open
Output dialog box as illustrated in Figure 1.53.
101
Figure 1.53 Open Output dialog box
Step 4. Look for the file name of the particular output that you want to
retrieve, Chapter 1 Output 1.spv in this case, and double-click on it; or
click it once and click Open or press Enter to get the needed output opened
immediately.
In case, you do not want to get to the Open Output dialog box from the
Window in the Menu Bar, Click File in the Menu Bar for the dropdown
dialog box of File to appear. In the File dropdown menu, select Open for
the Open dropdown dialog box to appear as presented in Figure 1.54; and
click Output…
102
Figure 1.54 Open dropdown dialog box in File, showing Output
With the click on Output…, the Open Output dialog box appears as
earlier shown in Figure 1.53. Double-click on the name of the output that
you want. If you do not want to double-click, just click on the output name
to highlight it, after which you click Open in the dialog box or press the
Enter key on the Keyboard.
103
Figure 1.55 File drop-down dialog box showing Close
When you get to closing of the last Data Editor file via the Close button
X or by clicking File and Exit, IBM SPSS prompts you that ‘Closing the
last Data Editor window will exit SPSS Statistics’ as shown in Figure
1.56; and asks you if you truly want to proceed with exiting the IBM SPSS
Statistics.
104
Figure 1.56 Prompt when exiting IBM SPSS Statistics
105
Chapter 2
DESCRIPTIVE STATISTICS
Overview
The endless quest for inventions, discoveries, new knowledge and theory
development, begins with questions that demand statistical descriptions as
the basis for solutions. Valuable information on the nature and
characteristics of variables in sampled groups are expressed descriptively.
The exquisite illustrations in this chapter guarantees every user the
attainment of expert knowledge of the various approaches in the utilization
of SPSS Statistics to practically analyse data for descriptive purposes.
106
problem and for better comprehension and utilization of other people’s
research works. Success in today’s world demands mastery and correct
application of certain statistical techniques as a precondition for one to
discover new knowledge and or apply existing knowledge most profitably.
It is with statistics that the necessary facts and information for correctly
answering research questions and testing the tenability of null hypotheses
are gathered. Without statistics, data collected in research cannot be
organized, analysed and synthesized objectively to arrive at principles,
paradigms, generalizations and theories for description, explanation,
prediction and control of future occurrence of specific phenomena.
Statistical application for right decision-making is at two levels: descriptive
and inferential. While descriptive statistics provides answers to research
questions, inferential statistics are used for testing of null hypotheses. See
Kpolovie (2018) for detailed account on parametric and nonparametric
inferential statistics.
Descriptive statistics that provide valuable information about the nature
and characteristics of individuals or other elements in a sampled group are
presented in this chapter. Measures of central tendency and variability are
thoroughly explained illustratively here.
Suppose that a public health worker was interested in examining 30-year
old single and married men’s attitude towards exclusive breastfeeding
(ATEB). She randomly drew a sample of 32 men who are at the age of 30
each and administered a 5-point scale (made up of 20 items, scored 1 for
strongly disagree through to 5 for strongly agree) on the variable to them;
and obtained the tabulated scores in Table 2.1. There are two variables of
interest in a study like this, dependent variable which is the ATEB and
independent variable which is marital status.
Note that the dependent variable, attitude towards exclusive
breastfeeding (ATEB), is a continuous variable that is measured at interval
scale of measurement such that a minimum possible score that a respondent
could get is 20 and a maximum possible score that a participant could
obtain is 100. Emphatically, ATEB which is the dependent variable is a
continuous variable and not a categorical variable. Counting and reporting
the number of respondents who answered strongly disagree, those who
opted for disagree, those who indicated undecided, those who responded
agree, and those who ticked strongly agree for each item on the scale is
inappropriate because such count can only demonstrate whether all the
107
scale points (1, 2, 3, 4 and 5) are selected with reasonable frequency;
thereby wrongly forcing the continuous variable to serve as a categorical
variable and totally defeating the purpose of the investigation. There is no
reason why a continuous variable like attitude towards exclusive
breastfeeding here should be treated as a categorical variable. A continuous
variable is capable of taking on large number of different values like
attitude towards exclusive breastfeeding in this case or scores on an
academic achievement test, intelligence quotient or a personality trait.
The independent variable, marital status, is a dichotomous categorical
variable that is measured at nominal scale of measurement, where arbitrary
numerical assignment of 0 and 1 or 1 and 2 is used merely as label for
differentiation. For no reason should a categorical variable like marital
status in this example be treated as a continuous variable. A categorical
variable is only capable of taking on highly limited number of different
values. When a categorical variable is limited to taking on only two values
like marital status in this case or gender, it is said to be a dichotomous
variable.
108
11 62 Single
12 26 Single
13 38 Single
14 40 Single
15 50 Single
16 66 Single
17 72 Married
18 35 Married
19 70 Married
20 68 Married
21 40 Married
22 50 Married
23 68 Married
24 49 Married
25 73 Married
26 67 Married
27 57 Married
28 74 Married
29 60 Married
30 65 Married
31 53 Married
32 60 Married
109
to IBM SPSS Statistics window appears as in Figure 2.1, simply close it by
clicking at top right corner of Welcome to IBM SPSS Statistics dialog
box, or by clicking Close at the bottom right corner and you will see the
IBM SPSS Statistics Data Editor (see Figure 2.2) for entering of data.
110
Step 3. Entering of data: Ensure that the cursor is in the extreme left and
topmost cell and click it (once clicked, the cell takes golden colour) so that
the data that you enter will begin from there. Very carefully type in the
values or scores listed under the Attitude towards exclusive breastfeeding
(ATEB) column in Table 2.1. Type each score and press the Enter key on
the Keyboard. Do the data entry without any mistake till you enter the last
value on ATEB. Then click the right-pointed arrow once and click the up-
pointed arrow on your Keyboard and hold it down until the cursor gets to
the topmost cell of that column (just next to the column that is now called
Var00001 where you have entered data on ATEB), and click it for the cell
to have golden colour as an evidence that the cursor is in it and that the
values you will enter now will begin from there.
The IBM SPSS Statistics Data Editor is set as default to accept only
numerical data. So, code every single man as 0 and every married man as
1 consistently. You could code every single man as 1 and married man as 2
consistently. Whichever code that you choose is only serving as numerals
assigned to dichotomously differentiate the 30-year old single (unmarried)
men from their counterparts who are married. To be on the same page, use
the code 0 for each single man and the code 1 for every married man that
you can see in column three of Table 2.1.
After carefully entering 0 for each single and 1 for each married man in
the second column of the Data Editor that is now called Var00002, the
next thing to do is to give the variables their actual names as treated in
Chapter 1. The entire process of naming variables is repeated here for you
to master. Try and master it as follows because it may not be repeated
subsequently. You will only be required to apply what you have learnt about
entering of data in Chapters 1 and 2 to data entry situations that you are
bound to encounter in every subsequent chapter.
Move cursor to and click Variable View at the bottom left corner of
the Data Editor to have the Variable View.
Put the cursor in the first cell under the Name column and click, then
type in the ATEB (spacing is not permitted) which is the actual name
for the dependent variable. On this same row, move the cursor to the
cell under Label, click it and type the dependent variable’s name in
full as Attitude towards exclusive breastfeeding. Also move cursor
to and click the cell under Measure column, click and select Scale
111
(meaning that the variable, ATEB, is measured at the interval scale of
measurement).
Move cursor to the second row under the Name column, click it and
type Maritalstatus (spacing is not allowed here please) which is the
name for the independent variable (i.e., the categorical variable). Still
on this same row, move the cursor to and click the cell under Label
and type Marital status (space is allowed here and is therefore used).
On this row also, move cursor to the column under Measure, click it
and select Nominal (meaning that Maritalstatus is a categorical
variable that is measured at nominal scale of measurement).
On the Maritalstatus row still, move cursor to the cell under Values
and click, after which you click inside the ellipsis button (i.e., the
small blue doted box) to have a dialog box. In this dialog box,
type 0 in the box beside Value and type Single in the box beside
Label as exemplified in Figure 2.3.
Then, click Add. Next, type 1 in the box beside Value and type
Married in the box beside Label as illustrated in Figure 2.4.
112
Figure 2.4 Value Labels for married
Finally, click OK to conclude the process for naming and labelling the
variables. Clicking the OK, takes you to the Variable View that has all the
variables’ names, labels and values as well as the measures (scales of
measurement) provided as in Figure 2.6.
113
Figure 2.7 Variable View with complete information
Take cursor to and click Data View at the very extreme bottom left of
the Variable View to see the Data Editor with all the data entered and
properly named. What you have on your screen will appear like the IBM
SPSS Statistics Data Editor in Figure 2.8.
114
115
Figure 2.8 IBM SPSS Statistics Data Editor with the ATEB data
Step 4. Save the entered data simply by moving cursor to the Toolbar and
clicking the tool for Save this document or simply press Crtl+S
(press and hold the control key [Ctrl] down and simultaneously tap S key
on the Keyboard). This action will produce the Save As… dialog box in
which you type a most suitable name for the dataset, Chapter 2 Table 1
Data.sav in this example, and click Save or press the Enter key on the
Keyboard as given in Figure 2.9. Ensure to save this working data as
Chapter 2 Table 1 Data.sav because it will subsequently be retrieved for a
number of statistical analyses.
Step 5. Analyze. As noted in Chapter 1, the fifth step, Analyze, in the use of
IBM SPSS Statistics for data analysis is where the entire statistical
operations for every statistical test is done. You are therefore required to
attend to everything that is done at the fifth step with unreserved
concentration. For each statistical test, a number of simple actions must be
taken to execute the analysis as explicitly illustrated. Note that for any
analysis to be done, the particular data set for the analysis must always be
activated (opened) in the IBM SPSS Statistics Data Editor.
116
Throughout the book, some situations where two or more selections or
clicking of statistical operations are to be done sequentially or successively,
I have used this rightward arrow ( ) to indicate the order of selection
117
Figure 2.10 Explore dialog box
118
Figure 2.11 Explore dialog box with variables moved to the right
119
Figure 2.13). Then, click Continue to return to the main Explore
dialog box.
120
Figure 2.14 Analyze, Descriptive Statistics and Explore
121
Figure 2.15 Explore dialog box with variable moved to the right
Click Continue. This takes you back to the main Explore dialog
box.
122
Finally, click OK in the Explore dialog box to have the entire
Descriptive Statistics Output as displayed in Output 2.1.
123
Histograms
124
Attitude towards exclusive breastfeeding Stem-and-Leaf Plot for
MaritalStatus= Single & Married
125
Men’s attitude towards exclusive breastfeeding
126
Attitude towards exclusive breastfeeding Stem-and-Leaf Plot
127
INTERPRETATION OF OUTPUT 2.1:
DESCRIPTIVE STATISTICS EXPLORE
Before continuing from Step 6, let me very briefly interpret the Output 2.1
of the Descriptive Statistics that has just been done. The first part of the
Output, termed Case Processing Summary, has shown that 16 Single and
16 Married men were the respondents and the Attitude towards exclusive
breastfeeding of each category of men was 100% valid as none of the
respondents had a missing score. While the percentage of valid cases for
each category was 100, the percentage of missing cases in single and in
married men was 0 (zero).
The second table of the Output, called Descriptives, has revealed that
single men’s attitude towards exclusive breastfeeding has a mean of
49.5625 with a Standard Error of 3.14108. It has a 95% Confidence limit of
128
42.8674 for Lower and 56.2576 for Upper, a Median of 50.5000, Variance
of 157.863, Standard Deviation of 12.56433, Minimum and Maximum
scores of 26 and 66, respectively, a Range of 40, and so on. Similar values
are equally given for the married men’s attitude towards exclusive
breastfeeding. Married men have a Mean of 60.0825, Standard Error of
2.96854 with 53.7352 and 66.3898 Lower and Upper bounds at 95%
certainty, a Median of 62.5000, Variance of 140.996, Standard Deviation of
11.87417, Minimum score of 35.00 and Maximum score of 74.00. Kurtosis,
Skewness and Interquartile Range are also given for each category of men.
Take very special note of the Mean and Standard Deviation as these are
the most important descriptive statistics that best capture the central
tendency and variability measures, and are therefore required for answering
research questions in both quantitative and qualitative research works.
The third part of the output (Attitude towards exclusive breastfeeding
histogram) are two histogram graphs plotted separately for attitude of
single men towards exclusive breastfeeding and of married men towards the
same dependent variable. Specifically, the first graph is for single men and
the second graph is for the married men.
Stem-and-Leaf Plots separately for single and married men’s attitude
towards exclusive breastfeeding constitute the fourth part of the output.
Frequency, stem and leaf as well as the stem width are given for single and
for married men. The two marital status (single and married) are plotted
against attitude towards exclusive breastfeeding.
The fifth part of the output gives the Case Processing Summary for all
the men, irrespective of marital status, who participated in the study. They
are 32 in number and all the 32 cases had valid score, making 100% valid
cases and 0.0% missing cases.
Descriptives is the sixth section of the output and it gives descriptive
statistical values for all the respondents taken together. The 32 men’s
attitude towards exclusive breastfeeding has a Mean of 54.8125, Standard
Error of 2.32554, Standard Deviation of 13.15525, Median of 57, Variance
of 173.060 with a Minimum score of 26 and Maximum score of 74. The
Lower and Upper bounds of the mean are respectively 50.0695 and 59.5555
at 95% Confidence Interval. The Range, Interquartile Range, Kurtosis and
Skewness of the distribution of the men’s scores are as well given.
A comprehensive Histogram of all the men’s scores is presented as the
seventh section of the Output. Lastly, Stem-and-Leaf Plots with Frequency,
129
Stem and Leaf and Stem width values are presented, ranging from about 28
to 74 with the centre lying between 45 and 65, and 58 as the core.
Step 6. Scroll up, down, left and right to view the Output on the screen of
your system with the aid of the up-pointed, down-pointed, left-pointed and
right-pointed arrows. These arrows are located at the right edge and bottom
edge of the IBM SPSS Statistics Viewer.
Step 7. Save the Output by clicking the Save this document tool in the
Toolbar to have its dialog box. Type the most suitable name for the output,
Chapter 2 Output 1.spv in this case, in the box beside File name and
press Enter or click Save as demonstrated in Figure 2.17 screenshot.
Step 8. Print the output and the input. When the Output is actively in view,
click the Print icon in the Toolbar to have its dialog box (see Figure
2.18) and click OK for one copy of the All visible output to be printed. For
printing of the data, activate the Data Editor by retrieving the data much
130
earlier saved as Chapter 2 Table 1 Data.sav. Click the Print icon in the
Toolbar. While viewing the data click OK in the Print dialog box for one
copy of the data to be printed.
Step 9. Exit IBM SPSS Statistics by closing all SPSS files that are open on
the computer screen. When you select Close for the last Data Editor file,
SPSS will prompt you that closing it will exit IBM SPSS Statistics and ask
you to confirm if really you want to exit SPSS (see Figure 2.18a). Click
Yes.
131
EXPLORE VIA SPSS SYNTAX DESCRIPTIVE
STATISTICS METHOD
Execution of Descriptive Statistics, like every other statistical analysis, can
with the greatest ease be done by applying IBM SPSS Statistics Syntax.
All that this method of performing Descriptive Statistics demands is
careful typing in of the right SPSS Syntax for the purpose. For execution of
the Descriptive Statistics of the data saved as Chapter 2 Table 1 Data.sav,
the IBM SPSS Syntax required to perform the same analysis without
wasting any time opening many dialog boxes and without clicking several
statistical operational options, is as shown in Syntax 2.1.
Once you enter this SPSS statistical Commands (Syntax) in the IBM
SPSS Statistics Syntax Editor, a single click of Run All will get the
entire analysis done and produce the results instantly, just like when the
analysis was executed by using SPSS Dialog Boxes Point and Click
132
technique. Mastering the IBM SPSS Statistics Syntax may be sounding
Greek to you and like a skill that cannot be very easily acquired.
Surprisingly, you will master it in no time with very little effort as the
whole IBM SPSS Statistics Commands for different statistical analysis are
made very handy in this book for your use. You only need to substitute the
variables names used in the examples with names of the variables in your
investigation. A quick trial will convince you. Just begin with this one and
congratulate yourself for the awesome success.
Click the file name, Chapter 2 Table 1 Data.sav, and press Enter or
click Open to open the document as earlier exhibited in Figure 2.8.
Open IBM SPSS Statistics Syntax Editor for entry of the SPSS
Syntax by clicking Window Go to Designated Syntax
133
Window for the Syntax Editor to open as illustrated in Figure 2.20.
You can equally have the IBM SPSS Statistics Syntax Editor opened
by clicking File New Syntax.
134
Click Run All. That is, click Run and in the Run dropdown
menu, click All as illustrated in Figure 2.22. Once you click All,
SPSS produces the Output in split seconds as presented in Output
2.2. Apart from clicking Run All to get the results produced,
you can simply select (highlight) the entire entered SPSS Syntax,
and click the Run Selection tool in the Toolbar. With a click
on the Run Selection tool, IBM SPSS Statistics instantly completes
the whole analysis and displays the results in the SPSS Viewer as
presented in Output 2.2.
135
Attitude towards exclusive breastfeeding
Histograms
136
137
Attitude towards exclusive breastfeeding Stem-and-Leaf Plot for
MaritalStatus= Single & Married
138
Men’s attitude towards exclusive breastfeeding
139
Attitude towards exclusive breastfeeding Stem-and-Leaf Plot
140
INTERPRETATION OF OUTPUT 2.2
Scroll through and view the entire Output 2.2 that was derived by analysing
the same data with the use of SPSS Syntax for Explore Descriptive
Statistics technique and notice that it is exactly equivalent to the Output
2.1, when the data were analysed, using SPSS Dialog Boxes Point and
Click method for the Descriptive Statistics via the Explore approach. Since
the Output 2.2 is parallel to Output 2.1, the same interpretation advanced
for the latter is equally appropriate for the former. So, do the interpretation
of Output 2.2 as was done for Output 2.1.
Alright, after interpreting the Output 2.2, complete the remaining steps
in data analysis with IBM SPSS Statistics thus
View output
141
Save the output
142
Start IBM SPSS Statistics
Retrieve the data since the entry has already been done and saved
143
Figure 2.24 Means dialog box with variables at the left
144
Figure 2.25 Means dialog box with variables moved to the right
145
Figure 2.26 Means: Options
In the Means: Options menu box, the default statistical operations that
will be performed are in the box at the right that is called Cell Statistics, all
the other additional statistical operations that can possibly be done with the
Means method are listed in the box at the left which is titled Statistics. You
are at liberty to choose which ones that you want to add to the three that
serve as the default. For purposes of illustration and for the fact that other
descriptive statistical processes produce many results, move all the
operations listed at the left into the right box to have a highly
comprehensive and very robust descriptive statistics done via the Means
method. To move a statistical operation in the left box into the right box,
highlight it and click the right-pointed arrow . But because you are
moving many (all of them in this example), from the beginning, Median at
the top, place the cursor, hold down Shift key and press and hold down the
arrow-down key on the Keyboard to highlight all of them. Then, click the
146
right-pointed arrow in the dialog box to get them moved at once into
the Cell Statistics box at the right as demonstrated in Figure 2.27. There is
nothing wrong if you prefer to move them one-by-one. Then click
Continue to get back to the Means main window.
Figure 2.27 Means: Options with all operations moved to the right
You do not need to worry at all about Style and Bootstrap in the Means
main menu box for the output to also indicate 95% Confidence bounds
lower and upper around the mean. But if you click Bootstrap in the Means
dialog box for instance, its window will appear (see Figure 2.28). In it,
checkmark the very small box beside Perform Bootstrapping; then you
may click the very small circle beside Bias accelerated (BCa) or leave it as
given by default to be Percentile. Be reminded that Bootstrap statistical
functions may most likely not be in lower Forms of SPSS (see Table 1.0)
like Base and Standard. Bootstrapping is used to indicate or test stability
and reliability of the mean for predictive modelling, particularly in
estimation of a parameter variance when standard assumptions about the
147
shape of the data are not met, which is not needed for the mere purpose of
descriptive statistics that is being done here.
After that, click Continue to return to the main Means window shown
in Figure 2.29.
148
Figure 2.29 Means window when it is fully ready for output production
149
INTERPRETATION OF OUTPUT 2.3
The Chapter 2 Output 3.spv in Output 2.3 has in the first part (Case
Processing Summary) shown that the scores of all the 32 men (100%)
were included in the analysis, meaning that no respondent’s score (0.0%)
was excluded. It can be discerned from the second part of the Output, called
Report, that the various descriptive statistical values are given separately
150
for Single, Married and Total, depicting attitudes toward exclusive
breastfeeding. For instance, the 16 single men have 49.5625 Mean,
12.56433 Standard Deviation, 50.5000 Median, 50.5000 Grouped Median,
3.14108 Standard Error of Mean, 793.00 Sum, 26.00 Minimum score, 66.00
Maximum score, 40.00 Range, 30.00 First entered score, 66.00 Last entered
score, 157.863 Variance, -.901 Kurtosis, 1.091 Standard Error of Kurtosis,
-.365 Skewness, .564 Standard Error of Skewness, 46.0672 Harmonic
Mean, 47.8976 Geometric Mean, which constitute 45.2% of the Total Sum
and 50.0% of the Total Number of cases.
The attitude of the 16 married men towards exclusive breastfeeding has
a Mean of 60.0625, Standard Deviation of 11.87417, Median of 62.5000,
Grouped Median of 63.3333, 2.96854 Standard Error of Mean, Sum of
961.00, Minimum score of 35, Maximum score of 74, Range of 39, 72.00
First entered score, 60.00 Last entered score, Variance of 140.996, Kurtosis
of -.246, 1.091 Standard Error of Kurtosis, -.799 Skewness, .564 Standard
Error of Skewness, 57.3820 Harmonic Mean, 58.8059 Geometric Mean,
constituting 54.8% of the Total Sum, and 50.0% of the Total Number of
respondents.
The entire 32 respondents as a group has 54.8125 Mean, 13.15525
Standard Deviation, 57.5000 Median, 57.5000 Grouped Median, 2.32554
Standard Error of Means, 1754.00 Sum, 26.00 Minimum score, 74.00
Maximum score, 48.00 Range, 30.00 First entered score, 60.00 Last entered
score, 173.060 Variance, -.713 Kurtosis, .809 Standard Error of Kurtosis,
-.469 Skewness, .414 Standard Error of Skewness, 51.1058 Harmonic
Mean, 53.0722 Geometric Mean that form 100% of Total Sum and 100% of
the Total Number of cases with regard to their attitude towards exclusive
breastfeeding. These values are the same with the descriptive statistics
values that were got when both SPSS dialog boxes point and click and
SPSS Syntax methods were employed in executing the analysis with
Explore.
Follow the remaining protocols to eventually shut down your system.
View output
151
Print output
152
Ensure that the data file, Chapter 2 Table 1 Data.sav, is currently open
on your computer screen. If it is not yet open, select File Open
Data Chapter 2 Table 1 Data.sav Open or Enter.
That is, click File, move cursor to Open, trace the cursor rightwards to
Data and click, look for and click the file named Chapter 2 Table 1
Data.sav, and click Open or press Enter. This action opens the data for
you to see on your screen as much earlier presented in Figure 2.8.
Simply click Window Go to Designated Syntax Window and
click (see Figure 2.30).
153
Finally, click Run and All to have the Output instantly in the SPSS
Viewer as demonstrated in Output 2.4.
Scroll through and view this Output 2.4 and compare it with the Output
2.3 when the data were analysed with Means dialog boxes point and click
method for Descriptive Statistics. You will quickly notice that the two
results are exactly the same in all the values. The Output 2.3 and Output 2.4
are very elaborate and robust that they do not need graphical illustration.
154
This accounts for why when descriptive statistics is done with the two
Means procedures, a graph does not constitute part of the output.
Adhere to the remaining protocols in data analysis with SPSS to
eventually shut down your system.
View output
Save the output as Chapter 2 Output 4.spv, and the Syntax as Chapter
2 Syntax 2.sps
Print output
155
A peculiar statistical function that the Descriptives method is uniquely
capable of, and which give the method a relative advantage over other
Descriptive Statistics procedures is the capacity to swiftly transform raw
scores to standard scores. With Descriptives method, to transform raw
scores observed into standardized scores, simply check or tick the Save
standardized values as variables checkbox in the process of the analysis.
This special feature is crucially needed in test development, validation and
standardization; and in the analysis of certain intelligence, achievement and
personality or affective tests. Transformation of raw scores to
standardized scores is indispensable for largescale testing. Detailed
attention is accorded it in Factor Analysis, Higher-Order Factor Analysis
and Multivariate Analysis that constitute a distinct book in the IBM SPSS
Statistics Excellent Guide series. Needless for much talk, execution of the
analysis is what matters most. So, get started.
Switch on the computer
Retrieve the data since the entry has already been done (but to have
better mastery of data entry, enter the data in Table 2.1, name and
label the variables, and save the file as earlier done in this chapter)
156
Figure 2.32 Analyze Descriptive Statistics Descriptives
157
variable from the left box into the Variable(s) panel at the right as
displayed in Figure 2.34.
158
Next, click the Options pushbutton to have its menu box in which you
ensure that the tiny box beside each of Mean, Sum, St. deviation, Minimum,
Maximum, Variance, Range, S.E. mean, Kurtosis and Skewness are ticked
as exemplified in Figure 2.35.
159
Saved standardized values
Respondents ATEB ZATEB
1 30.00 -1.88613
2 40.00 -1.12598
3 55.00 .01425
4 51.00 -.28981
5 45.00 -.74590
6 58.00 .24230
160
7 60.00 .39433
8 62.00 .54636
9 44.00 -.82192
10 66.00 .85042
11 62.00 .54636
12 26.00 -2.19019
13 38.00 -1.27801
14 40.00 -1.12598
15 50.00 -.36582
16 66.00 .85042
17 72.00 1.30651
18 35.00 -1.50605
19 70.00 1.15448
20 68.00 1.00245
21 40.00 -1.12598
22 50.00 -.36582
23 68.00 1.00245
24 49.00 -.44184
25 73.00 1.38253
26 67.00 .92644
27 57.00 .16628
28 74.00 1.45854
29 60.00 .39433
30 65.00 .77441
31 53.00 -.13778
32 60.00 .39433
161
INTERPRETATION OF OUTPUT 2.5
It can be seen from the first part of Output 2.5 (Descriptives) that the 32
respondents’ attitude towards exclusive breastfeeding has a Range Statistic
of 48.00, Minimum Statistic of 26.00, Maximum Statistic of 74.00, Sum
Statistic of 1754.00, Mean Statistic of 54.8125, Mean Standard error of
2.32554, Standard Deviation Statistic of 13.15525, Variance Statistic of
173.060, Skewness Statistic of -.469, Skewness Standard error of .414;
-.713 Kurtosis Statistic and .809 Kurtosis Standard error. These computed
values are the same as the ones got when the data were analysed with the
two Explore methods and two Means methods.
The second part of the Output, called Saved standardized values, that
is critically important, and is completely unique to the Descriptives method;
has shown z Score for each of the 32 men’s attitude towards exclusive
breastfeeding (ZATEB). For instance, while the first man with a raw score
of 30 has a z Score of -1.88613, the second who scored 40 has a z Score of
-1.12598, the third with a raw score of 55 has a z Score of .01425, the
seventeenth respondent with a raw score of 72 has a z Score of 1.30651, the
thirtieth participant with a raw score of 65 has a z Score of .77441, and so
on. This type of descriptive information is only possible with the
Descriptives technique for establishing descriptive statistics. Note that z
Score is a standardized score with a mean of 0 and a standard deviation of
1. Also, that the z Scores are not produced directly in the IBM SPSS
Statistics Viewer window, though it is part of the output, but in the IBM
SPSS Statistics Data Editor as an additional column.
162
Syntax 2.3 Chapter 2 Syntax 3.sps
Ensure that the data file, Chapter 2 Table 1 Data.sav, is currently open
on your computer screen. If it is not yet open, select File Open
Data Chapter 2 Table 1 Data.sav Open or Enter.
That is, click File, move cursor to Open, trace the cursor rightwards to
Data and click, look for and click the file which is named Chapter 2 Table
1 Data.sav, and click Open or press Enter. Now, follow these five simple
steps.
i. Select Window.
ii. Select Go to Designated Syntax Window.
iii. Enter the Syntax 2.3.
iv. Select Run.
v. Select All.
That is, select Window and in its dropdown menu, click Go to
Designated Syntax Window to have the Syntax Editor where you
carefully type in the Syntax 2.3. When done, the IBM SPSS Syntax Editor
will be like Figure 2.36.
163
Finally, click Run All for the results to instantly be displayed in
the IBM SPSS Statistic Viewer as in Output 2.6.
164
7 60.00 .39433
8 62.00 .54636
9 44.00 -.82192
10 66.00 .85042
11 62.00 .54636
12 26.00 -2.19019
13 38.00 -1.27801
14 40.00 -1.12598
15 50.00 -.36582
16 66.00 .85042
17 72.00 1.30651
18 35.00 -1.50605
19 70.00 1.15448
20 68.00 1.00245
21 40.00 -1.12598
22 50.00 -.36582
23 68.00 1.00245
24 49.00 -.44184
25 73.00 1.38253
26 67.00 .92644
27 57.00 .16628
28 74.00 1.45854
29 60.00 .39433
30 65.00 .77441
31 53.00 -.13778
32 60.00 .39433
165
The results of the analysis, using IBM SPSS Syntax Descriptives
method for the Descriptive Statistics execution, are exactly like the ones
that were arrived at via the SPSS dialog boxes point and click method that
have been presented earlier in Output 2.5.
Follow the remaining SPSS Statistics data analysis protocols of
View output
Print output
166
data must have been collected at nominal and or ordinal scale of
measurement that can only accept frequency of cases classified into each
category without the possibility of statistical operations that require
addition, subtraction, division and multiplication.
At the second level, Frequencies method provides detailed measures of
central tendency and measures of variability or dispersion for each of the
dependent quantitative or continuous variables. Data at the second level of
Frequencies statistics must have been collected with summative response
at interval and or ratio scales of measurement that can accept all arithmetic
operations of addition, subtraction, division and multiplication as done
when calculating mean and standard deviation for instance. The
Frequencies procedure can also provide graphs like histogram, pie chart
and bar chart for each dependent variable in line with the investigator’s
choice.
The greatest strength of the Frequencies method of Descriptive
Statistics analysis is in allowing for comparison of two or more quantitative
or continuous variables simultaneously. Apart from that, the Frequencies
approach is not suitable for comparing a dependent variable separately in
line with two or more categories of an independent variable. In the working
example for instance, Frequencies cannot provide descriptive statistics
separately for the single and married categories of men, as it can only give
the descriptive statistics for both groups taken together. In fact,
Frequencies technique provides descriptive statistics more like the example
of the quick descriptive statistics that was done in Chapter 1 by merely
clicking the Run descriptive statistics tool in the Toolbar. Recall that
the output of descriptive statistics done in Chapter 1 via clicking the Run
descriptive statistics tool merely provided the descriptive statistical values
for the two groups jointly on the dependent variable. It is that type of output
that the Frequencies method is capable of producing. Enough of the
narrative, commence the analysis itself by adopting the ten steps for
application of IBM SPSS Statistics in data analysis.
167
Step 3. Enter the data in Table 2.1, name and label the variables. The said
data have been entered and saved earlier as Chapter 2 Table 1 Data.sav.
So, you may very quickly retrieve it by moving the cursor to the Menu Bar
in the Data Editor and selecting File Open Data
Chapter 2 Table 1 Data.sav Open or Enter as depicted in Figure
2.37 and Figure 2.38.
168
Figure 2.38 Open Data Chapter 2 Table 1 Data.sav dialog box
In the Open Data dialog box, select the file name, Chapter 2 Table 1
Data.sav Open or Enter. That is, select the file name and click
Open or press Enter to have the said file opened on the screen.
Alternatively, instead of selecting File Open Data
Chapter 2 Table 1 Data.sav Open or Enter as shown in Figures
2.37 and 2.38 in order to open the needed data file, from the IBM SPSS
Statistics Data Editor very quickly click Chapter 2 Table 1
Data.sav Open or Enter to have the file opened. This means that in
the Data Editor, click the Open data document icon in the Toolbar,
then click the file name in the Open Data dialog box that appeared as in
Figure 2.38 and click Open in it or press Enter on the Keyboard for the
needed file to be opened. In other words, while in Data Editor, merely
clicking automatically opens the dialog box in Figure 2.38 where you
169
select and open the particular data document that you want (in this case,
Chapter 2 Table 1 Data.sav) as in Figure 2.8.
Step 4. Save the data if you entered the data afresh. But if you retrieved the
dataset, this Data file has since been saved, so go on to the next step where
the specific statistical operations for the Frequencies analysis will be
performed.
170
Figure 2.40 Frequencies dialog box
This dialog box (Figure 2.40) is used to perform the statistical operations
required for execution of the first level of Frequencies descriptive
statistics that provides frequency counts for the different categories of each
of the independent variables that are categorical in nature. To get this done,
highlight the independent variable (Marital status in this case) that is in the
box at the left and move it to the box under Variable(s) that is at the right
by clicking the right-pointed arrow that is directly facing the
Variable(s) panel at the right. Once done, the Marital status vanishes from
the left box and appears in the Variable(s) box at the right as shown in
Figure 2.41.
171
Figure 2.41 Frequencies dialog box with the Marital status moved
172
Figure 2.42 Open Data menu box
173
Figure 2.43 Analyze Descriptive Statistics Frequencies
Once you click Frequencies, its menu box will appear as in Figure 2.44.
Move cursor to, and click Reset to terminate previous statistical analysis
operations done. The termination does not affect the already pasted results.
By clicking Reset, you must have observed that the independent variable
(Marital status) automatically disappeared from the Variable(s) box at the
right box and appeared in the left box as in Figure 2.45.
174
Figure 2.45 Frequencies menu box after clicking Reset
175
Click Statistics pushbutton to have the Frequencies: Statistics dialog
box in which the various statistical operations like measures of Central
Tendency, measures of Dispersion (variability), Percentile Values and
Distributions other than normal distribution are listed as depicted in
Figure 2.47.
176
checked, the Frequencies Statistics dialog box will look like what is shown
in Figure 2.48.
Now, move cursor to and click Continue, and it will take you back to
the main Frequencies menu box (see Figure 2.49). Note that at this time,
the entire Frequencies statistical operations you want to do for the
dependent variable (Attitude towards exclusive breastfeeding) are very
ready to be displayed apart from graphic presentation.
177
Figure 2.49 Frequencies menu box to opt for Charts
In order to have either a Bar chart, Pie chart or Histogram for each
dependent variable as part of the results, click Charts pushbutton for its
dialog box to appear as shown in Figure 2.50.
In the Frequencies: Charts dialog box, click Histogram and the circle
beside it gets shaded. Because it is good to have a rough idea of how much
the Histogram emanating from the dependent variable approximates a
normal curve, also tick the checkbox beside Show normal curve on
178
histogram as indicated in Figure 2.51. Next, click Continue to get back to
the Frequencies main menu when all the analysis needed have been done.
179
Attitude towards exclusive breastfeeding
Cumula
Frequency Percent Valid Percent tive %
Valid 26 1 3.1 3.1 3.1
180
30 1 3.1 3.1 6.3
35 1 3.1 3.1 9.4
38 1 3.1 3.1 12.5
40 3 9.4 9.4 21.9
44 1 3.1 3.1 25.0
45 1 3.1 3.1 28.1
49 1 3.1 3.1 31.3
50 2 6.3 6.3 37.5
51 1 3.1 3.1 40.6
53 1 3.1 3.1 43.8
55 1 3.1 3.1 46.9
57 1 3.1 3.1 50.0
58 1 3.1 3.1 53.1
60 3 9.4 9.4 62.5
62 2 6.3 6.3 68.8
65 1 3.1 3.1 71.9
66 2 6.3 6.3 78.1
67 1 3.1 3.1 81.3
68 2 6.3 6.3 87.5
70 1 3.1 3.1 90.6
72 1 3.1 3.1 93.8
73 1 3.1 3.1 96.9
74 1 3.1 3.1 100.0
Total 32 100.0 100.0
181
INTERPRETATION OF OUTPUT 2.7
The Output 2.7 has two parts in accordance with the two levels of
Frequencies Descriptive Statistics. At the first level, the results have
shown Frequencies Statistics for the independent variable, marital status,
that has two categories (single and married). The Statistics section here
indicates 32 valid cases and that no single case is missing. The sub-table
called Marital status gives frequency counts of 16 for married and 16 for
single, with a total of 32 men. While 50% of the cases are single, the other
50% of cases are married. The total is 100%. Similarly, the valid percent is
50 for each of the two categories, giving a total of 100% valid cases of the
respondents.
The second part of the Output 2.7 that is called Frequencies Statistics
of the dependent variable, Attitude towards exclusive breastfeeding, has
32 valid scores (none missing), 54.8125 Mean, 2.32554 Standard Error of
Mean, 57.5000 Median, 40.00 Mode, 13.15525 Standard Deviation,
173.060 Variance, -.469 Skewness, .414 Standard error of Skewness, -.713
182
Kurtosis, .809 Standard Error of Kurtosis, 48.00 Range, 26.00 Minimum
score, 74.00 Maximum score, 1754.00 Sum, 44.2500 for 25 Percentiles,
57.500 for 50 percentiles and 66.0000 for 75 percentiles.
The second table titled Attitude towards exclusive breastfeeding has
listed the different scores on the dependent variable and provided for each
score the frequency of occurrence, percentage of occurrence, indicating
both the valid percentage and the cumulative percentage. For instance, the
score 74 occurred once, which is 3.1 percent and 3.1 valid percent of total
scores, and it has a cumulative percent of 100.
Lastly, the attitude of the 32 men towards exclusive breastfeeding is
plotted with a histogram graphically. To provide an idea of how the men’s
attitude approximates normal distribution, a normal curve is superimposed
on the histogram bars. The scores do not depart too much from the normal
curve.
Haven completed the analysis as demanded by Analyze (the Step 5) of
application of IBM SPSS Statistics in execution of data analysis, the
remaining protocols to adhere are shown from Step 6 to Step 10:
View output
Step 6. Scroll up and down, and possibly left and right to view the entire
output on the screen of your system with the aid of the up-pointed, down-
pointed, left-pointed and right-pointed arrows. These arrows are located at
the right edge and bottom edge of the screen.
183
Step 7. Save the Output by clicking the Save this document icon in
the Toolbar to have its dialog box. Type the most suitable name for the
output, Chapter 2 Output 7.spv in this case, in the box beside File name
and press the Enter key on the Keyboard or click Save in the Save Output
As dialog box (see Figure 2.52).
Step 8. Print the output and the input. When the output is actively in view,
click the Print icon in the Toolbar to have its dialog box, shown in
Figure 2.53 and click OK for the output to be printed. For printing of the
data, activate the SPSS Date Editor by retrieving the data much earlier
184
Figure 2.53 Print output dialog box
Step 9. Exit IBM SPSS Statistics by closing all SPSS files that are open on
the computer screen. When you select Close for the last Data Editor file,
SPSS will prompt you that closing it will exit IBM SPSS Statistics, and
ask you to confirm if really you want to exit SPSS as given in Figure 2.54.
Click Yes.
185
Switch on the computer
Retrieve the data since the entry has already been done and saved as
Chapter 2 Table 1 Data.sav
186
Figure 2.55 Open Data menu box
187
With the data file open as in Figure 2.8, to activate the IBM SPSS
Statistics Syntax Editor while in the SPSS Data Editor that contains the
dataset, click Window Go to Designated Syntax Window and click
as illustrated in Figure 2.56.
188
Figure 2.57 Syntax window with Frequencies Syntax rightly entered
Finally, click Run All to have SPSS produce the results in the
SPSS Viewer instantly as displayed on your screen and demonstrated in
Output 2.8.
189
Attitude towards exclusive breastfeeding
Frequency Percent Valid Percent Cumulative Percent
Valid 26 1 3.1 3.1 3.1
30 1 3.1 3.1 6.3
190
35 1 3.1 3.1 9.4
38 1 3.1 3.1 12.5
40 3 9.4 9.4 21.9
44 1 3.1 3.1 25.0
45 1 3.1 3.1 28.1
49 1 3.1 3.1 31.3
50 2 6.3 6.3 37.5
51 1 3.1 3.1 40.6
53 1 3.1 3.1 43.8
55 1 3.1 3.1 46.9
57 1 3.1 3.1 50.0
58 1 3.1 3.1 53.1
60 3 9.4 9.4 62.5
62 2 6.3 6.3 68.8
65 1 3.1 3.1 71.9
66 2 6.3 6.3 78.1
67 1 3.1 3.1 81.3
68 2 6.3 6.3 87.5
70 1 3.1 3.1 90.6
72 1 3.1 3.1 93.8
73 1 3.1 3.1 96.9
74 1 3.1 3.1 100.0
Total 32 100.0 100.0
191
It can be discerned that the Output 2.8 for when the descriptive statistics
was done with Frequencies SPSS Syntax method is exactly the same with
Output 2.7 when the Frequencies SPSS dialog boxes point and click
technique was used for execution of the descriptive statistics. With the two
outputs being equivalent, the same explanation of the Output 2.7 earlier
given, applies here too for Output 2.8.
Follow the remaining basic protocols for data analysis with IBM SPSS
Statistics to eventually shut down your system.
View output
Save the output as Chapter 2 Output 8.spv, and the Syntax as Chapter
2 Syntax 4.sps
Print output
192
Exit IBM SPSS
193
Chapter 3
INDEPENDENT-SAMPLES t TEST
Overview
The most appropriate inferential statistical technique for comparison of two
groups with respect to a specific dependent variable is the Independent-
Samples t Test. The assumptions that underlie comparison of means are
elucidated. How to personally utilize SPSS Statistics in the analysis of data
collected independently from two groups for accurate comparison of the
populations that the groups represent is presented in this chapter for every
user to effortlessly master. Dialog boxes selection and syntax methods are
excellently illustrated with screenshots. Gender difference in basic science
concepts, and attitude towards exclusive breastfeeding by marital status are
used for demonstration.
194
adopted for determination of statistically significant difference between two
independent samples with respect to a particular dependent variable and a
dichotomous categorical or grouping independent variable. One-sample t
Test is applied for comparison of a sample mean with a known or estimated
population mean. Paired-samples t Test is used for comparison of two
different means from correlated, dependent or paired samples. Whenever
two means obtained by the same subjects are to be compared, paired-
samples t Test is utilized for the purpose.
195
Independence of observations
Application of t Test assumes that the different populations from which
samples are to be drawn are completely independent of each other and that
the samples are randomly drawn from the populations in a completely
independent manner. That is, the random selection of each of the
individuals or elements from population A to constitute sample A is totally
independent of the random selection of each of the subjects from population
B to form sample B. Independence of observations tends to be the most
important of the assumptions that underlie comparison of means statistical
techniques and other inferential statistical approaches because violation of
this assumption is capable of causing serious inferential errors. Indeed, the
best strategy for achieving total independence in research is by explicit
randomization that ensures random drawing of general sample and random
assignment of subjects as well as treatment conditions to the required
groups (Kpolovie, 2018; 2016; 2011; 2010).
Homoscedasticity of variance
Mean comparisons as in t Test always assumes homoscedasticity which is
homogeneity of variance or equality of variance in the different populations
from which random samples are taken for the study. The variance of the
dependent variable should be comparable across levels of the independent
variable. In t Test, homoscedasticity or homogeneity of variance assumption
is addressed by a default in the IBM SPSS Statistics. Apart from that, to
ensure homogeneity of variance, I very strongly recommend that sample
sizes of the various groups should be equal or at least approximately equal.
With equal sample sizes that are each suitably large, homoscedasticity
assumption may be violated without serious risk (Kpolovie, 2018).
196
Concepts to 69 randomly sampled K-6 students, and they obtained the
scores listed in Table 3.1. (You may recall that the scores of the first 20
participants here were used for illustration in Table 1.1 of Chapter 1).
197
22 Male 45 57 Female 46
23 Female 40 58 Female 42
24 Male 35 59 Male 40
25 Female 50 60 Male 41
26 Male 30 61 Female 43
27 Female 45 62 Male 43
28 Male 30 63 Female 40
29 Female 40 64 Female 39
30 Male 25 65 Female 39
31 Male 50 66 Male 52
32 Male 45 67 Male 39
33 Female 55 68 Male 46
34 Male 55 69 female 56
35 Male 50
198
EXECUTION OF THE INDEPENDENT-SAMPLES t
TEST ON BSCS
Follow the ten steps for performing data analysis with IBM SPSS Statistics.
Step 2. Launch IBM SPSS Statistics by clicking its icon . Wait for it to
completely get ready. That is, till when the Welcome to IBM SPSS
Statistics shows as illustrated in Figure 3.1.
Close the superimposed dialog box and be in the IBM SPSS Statistics
Data Editor. It is in this Data Editor that entry of the data is done.
Step 3. Enter the data very carefully in the Data View of the IBM SPSS
Statistics Data Editor. After the data entry, switch to the Variable View by
clicking it. In the Variable View, type in the Name, Label, Values, and
select the Measure for the dependent variable (Basic Science Concepts –
BSCS) and for the independent variable (Gender) as earlier exemplified in
Chapter 1. When you are done, click Data View to return to the Data View
that will be as shown in Figure 3.2.
199
200
Figure 3 Continued
201
Figure 3.2 Continued
Step 4. Save the data by moving cursor to the Toolbar and clicking Save
this document icon . Alternatively, move cursor to the Menu Bar and
click File for its dropdown dialog box in which you click Save As… for the
Save Data As dialog box to appear. Then, type a suitable file name
202
(Chapter 3 Table 1 Data.sav in this case) in the File name field as shown
in Figure 3.3. After that, click Save or press Enter.
Figure 3.3 Save Data As… dialog box with the file name
203
Figure 3.4 Analyze Compare Means Independent-Samples T Test
204
Move the Basic Science Concepts from the left box into the Test
Variable(s) panel at the right by highlighting it and clicking the right-
pointed arrow that is directly facing the Test Variable(s) panel. Also,
highlight the Gender in the left box and move it to the Grouping Variable
panel at the right by clicking the right-pointed arrow that is directly
facing the Grouping Variable box. With the dependent variable (Basic
Science Concept) and the independent variable (Gender) moved into their
designated panels at the right, the Independent-Samples T Test menu
takes the form that is presented in Figure 3.6.
205
Figure 3.7 Independent-Samples T Test menu box requesting you to Define Groups…
Were you able to notice any difference? Yes, certainly. Define Groups
has now been activated fully. Click the Define Groups… to have its dialog
box in which you type 1, the code for Male, in the Group 1 panel; and type
2, the code for Female, in the Group 2 panel as exhibited in Figure 3.8.
Figure 3.8 Define Groups dialog box with the Groups defined
206
Figure 3.9 Independent-Samples main menu
In case you want to see and perhaps change how IBM SPSS Statistics
intends to treat the percentage of Confidence Interval and Missing Values
in the analysis, click Options pushbutton. SPSS sets 95% Confidence
Interval and Exclude cases analysis by analysis as the default for
performing the analysis. See the screenshot indicating the IBM SPSS
Statistics default for treating Confidence Interval Percentage and Missing
Values in Figure 3.10. Then, click Continue to go back to the
Independent-Samples T Test main menu as presented in Figure 3.9.
207
the dependent variable’s curve or distribution. In the Bootstrap dialog box,
check Perform bootstrapping checkbox as well as the Bias corrected
accelerated (BCa) checkbox. These Bootstrapping operations are actually
not needed in the execution of the Independent-Samples t Test that is under
consideration in this chapter which has from the onset, met the assumptions
required for comparison of means with parametric statistical tests.
Bootstrap is a nonparametric procedure for arriving at statistical inference
when the data do not meet the assumptions about underlying population
distributions.
208
OK will get IBM SPSS Statistics to instantly generate the output. So,
move the cursor to and click OK for the results to be presented in the IBM
SPSS Statistics Viewer as demonstrated in Output 3.1.
209
is not significant as its Sig. of .676 is greater than the chosen alpha of .05.
Therefore, the t Test results in the row of Equal variances assumed are the
ones that are to be correctly reported. If the Levene’s test had been
significant, the results in the row of Equal variances not assumed could
have been reported instead. The computed t is -1.087 with 67 degrees of
freedom, .281 Sig. (2-tailed), -2.95882 Mean Difference, 2.72093
Standard Error Difference plus -8.38983 and 2.47219 respectively for
Lower and Upper 95% Confidence Interval of the Difference. The
calculated t of -1.087 with -2.95882 mean difference is not significant
because the probability (Sig.) of obtaining this result is .281 that is greater
than the chosen alpha of .05. Therefore, the null hypothesis postulated
much earlier that “There is no significant difference in K-6 students’ Basic
Science Concepts by gender” is retained; t(67) = -1.087, p > .05, 2-tailed.
Females (mean = 39.5588 and standard deviation = 11.50017) do not
perform significantly better than males (mean = 36.6000 and standard
deviation = 11.10167) in Basic Science Concepts.
The Mean Difference of -2.95882 between male and female students’
Basic Science Concepts is rather a function of chance because it is not
statistically significant at the chosen alpha of .05 or at the 95% certainty. In
research, a difference is indeed a difference only when it is statistically
significant, when replication of the work 100 times with other samples
randomly drawn from the same population will not produce different results
more than 5 times (Kpolovie, 2018).
It must finally be emphasized that whenever the Lower and Upper 95%
Confidence Interval of the Difference both fall either above zero or below
zero, the pair of means under comparison statistically differ significantly.
That is, the mean difference between the two means is significant. But in
the analysis that the output is being interpreted here, the Lower and the
Upper 95% Confidence Interval of the Difference do not both fall either
above or below zero (-8.38983 [below zero] and 2.47219 [above zero],
respectively). Therefore, the Mean Difference of -2.95882 is not statistically
significant and its effect size is bound to be “trivial” or at best, “small” as
will soon be demonstrated. In other words, whenever the Lower 95%
Confidence Interval of the Difference falls below zero and the Upper 95%
Confidence Interval of the Difference falls above zero, the pair of means
under comparison do not significantly vary. The mean difference in the
scenario that zero lies between the Lower and Upper 95% Confidence
210
Interval of the Difference is not significant and is totally a mere function of
chance.
211
Table 3.2 Recommendations for classification of d as effect size
d AS THE EFFECT DESCRIPTIVE CLASSIFI-
SIZE IN t Test CATION OF THE d
0.000 to 0.199 Trivial effect size
0.200 to 0.499 Small effect size
0.500 to 0.799 Medium effect size
0.800 and above Large effect size
Where M1 is the mean for the first group, M2 is the mean for the second
group, and Spooled is the pooled standard deviations of the two groups. The
numerator can either be M1 – M2 or M2 – M1 consistently. Just be sure to
correctly report the side (group) that the mean difference, if any, favours. If
the population standard deviation is known, the Cohen’s Effect Size statistic
(d) is simply got by dividing the mean difference between the two groups
212
by the population standard deviation (M1—M2/σ directly instead of
estimating the population standard deviation with the Spooled. Since in most
research settings, the population parameters are not known, the Spooled must
be used to estimate the population standard deviation. The Spooled is the best
estimate of the population standard deviation with respect to the construct
under investigation. The Spooled is computed as:
213
This computed d (-.262) is rightly interpreted as “small effect size” of
the independent variable on the dependent variable. That is, students’
Gender accounts for only small percent of the variance in the students’
Basic Science Concepts in favour of the female students. Recall that the
computed t value of -1.087 has since been found to be insignificant.
The entire results can therefore be summarized as thus. The basic
science concepts scores of male students (M = 36.6000; SD = 11.10167)
do not significantly differ from the basic science concepts scores of the
female students (M = 39.5588; SD = 11.50017): t(67) = -1.087; p > .05
(two-tailed); d = -.262 (a small effect size). Therefore, there is no
214
significant difference in K-6 students’ basic science concepts with regard to
gender.
In a scenario with equal sample size for the two groups, the d can also be
computed simply with Equation 3:4, which states that the Cohen’s d as a
measure of Effect Size for Independent-Samples t Test is the mean
difference between the two groups divided by the average of the square root
of the variances of both groups. That is:
Where M1 and M2 are the means for the first group and the second
group, respectively; S12 and S22 are the variances (square of the standard
deviations) for the two respective groups. Assuming that the two groups
(male and female students) in the current example are equal in size, the d
can merely be computed as:
215
The d got here with Equation 3:4 is the same as the d got with Equation
3:3 because the two groups are almost exactly equal in size (35 and 34).
This is in line with my recommendation under homoscedasticity assumption
that ‘to ensure homogeneity of variance, I very strongly recommend that
sample sizes of the various groups should be equal or at least approximately
equal. With equal sample sizes that are each suitably large,
homoscedasticity assumption may be violated without serious risk.’
Apart from Equation 3:3 and Equation 3:4, irrespectively of whether the
two groups are equal in sample size or not, Cohen’s d as a measure of Effect
Size for Independent-Samples t Test can be computed directly with the
calculated Independent-Samples t Test value, using Equation 3:5. With
Equation 3:5, the Effect Size for Independent-samples t Test is:
Where N1 is the sample size of the group 1; N2 is the sample size of the
group 2; and t is the computed Independent-Samples t Test. In the current
example that the N1 is 35; N2 is 34; and t is -1.087, the d can be got as:
The d of -.262 has already been interpreted as a “small effect size” of the
grouping variable on the dependent variable. It must be noted that the minus
sign (“-“) in the d does not denote negativity or “less than 0” literally. It
rather reflects the side or group that the mean difference favours. The M1
(36.6000) for male students is smaller than the M2 (39.5588) for the female
216
students. If in computing the d, the male’s mean (M1) is subtracted from the
mean of the female students (M2), the result will simply be .262, still
denoting that the Basic Science Concepts mean of the female students is a
little higher than that of their male counterparts. Thus, the computed -.262 d
does not mean less than zero. Be again reminded that the mean difference,
as revealed by the t, is not significant statistically and so, denotes chance
occurrence. Haven completed interpretation of the Output 3.1, follow the
remaining five steps in data analysis with SPSS Statistics to shut down the
computer.
Step 6. View the output by scrolling up, down, right and left to see every bit
of it.
217
Step 8. Print the Output by merely clicking the Print icon in the
Toolbar. If not so, click File and from the File dropdown dialog box, click
Print while in the Viewer window with the Output 3.1 showing on the
screen to have the Print Output menu as given in Figure 3.13.
Step 9. Exit IBM SPSS Statistics by closing all the SPSS files that are open.
When on clicking the last Data Editor to close it, IBM SPSS Statistics
prompts you to confirm that indeed you want to exit SPSS, proceed and
confirm it by clicking Yes.
Step 10. Shut down the computer. Simply select Start Power
Shut down.
218
SPSS SYNTAX FOR INDEPENDENT-SAMPLES t
TEST ON STUDENTS’ BSCS
The IBM SPSS Statistics Syntax with which to perform the Independent-
Samples t Test on the K-6 students’ Basic Science Concepts by Gender is
as shown in Syntax 3.1.
Steps 3 and 4. Retrieve the data since the data are already entered, named,
and saved. In the Data Editor, move cursor to the Toolbar and click Open
data document tool or icon to have the Open Data dialog box (see
Figure 3.15). Alternatively, move cursor to the Manu Bar and click File
Open Data (see Figure 3.14) to have the Open Data dialog
box shown in Figure 3.15.
219
Figure 3.14 File Open Data
Look for the name (Chapter 3 Table 1 Data.sav) with which the data
file was saved, click it to highlight it and click Open or press Enter on the
220
Keyboard. With this, the Chapter 3 Table 1 Data.sav opens as presented in
Figure 3.2.
Step 5. Analyze the data with the IBM SPSS Statistics Syntax method by:
Activate or open the IBM SPSS Statistics Syntax Editor. To do
this, move cursor to the Menu Bar and click Window Go to
Designated Syntax Window as shown in Figure 3.16 for the Syntax
Editor to open. Once the Go to Designated Syntax Widow is
clicked, the SPSS Syntax Editor appears on your screen as shown in
Figure 3.17. The IBM SPSS Statistics Syntax Editor can also be
opened by clicking File New Syntax. That is, click File
and in the File dropdown menu, select New; and in the dropdown of
the New, click Syntax.
221
Figure 3.18 SPSS Syntax Editor with the Syntax entered
222
INTERPRETATION OF OUTPUT 3.2
It is glaring that the Output 3.1 got through dialog boxes point and click
method and the Output 3.2 arrived at through SPSS Syntax technique for
execution of the Independent-Samples t Test are exactly the same. Go
through the interpretation of Output 3.1 again and very carefully write out
an interpretation of the Output 3.2 accordingly.
Learning to apply SPSS Syntax is, after all, very easy and very fast. It is
like a short cut to execution of statistical analysis with the use of IBM SPSS
Statistics. The use of SPSS Statistics Syntax is, therefore, very much worth
mastering. Do not miss it.
Step 6. Scroll with the aid of the right-pointed, left-pointed, up-pointed and
down-pointed arrows at the right edge and bottom edge of the computer
screen to view the output.
Step 7. Save the output as Chapter 3 Output 2.spv and the Syntax as
Chapter 3 Syntax 1.sps. In the Viewer window, move cursor to the
223
Toolbar and click Save this document icon and in the box beside File
name, type in the name Chapter 3 Output 2.spv and click Save or press
Enter (see Figure 3.20).
In like manner, be in the Syntax Editor with the entered Syntax and
click Save this document icon in the Toolbar and type in the File
name as Chapter 3 Syntax 1.sps and click Save or press Enter to get the
Syntax saved.
Step 8. Print the output and the Syntax files just the way such files should
be printed as discussed earlier.
Step 9. Close all opened SPSS files and Exit IBM SPSS Statistics.
Step 10. Click Start Power Shut down to shut the system
down.
224
DIALOG BOXES SELECTION FOR
INDEPENDENT-SAMPLES t TEST ON EXCLUSIVE
BREASTFEEDING
Acquisition of expert knowledge of IBM SPSS Statistics in data analysis
that this book guarantees users can most easily be achieved with relentless
practice. The data in Chapter 2 on men’s attitude towards exclusive
breastfeeding with regard to marital status shall very quickly be used here
to further ensure your mastery of Independent-Samples t Test execution
with IBM SPSS Statistics. The investigator must have asked a research
question and postulated both alternate and null hypotheses thus.
Research question
What is the attitude of single and married men towards exclusive
breastfeeding as measured by their mean and standard deviation?
Alternate hypothesis
There is a statistically significant difference in the attitude of single and
married men towards exclusive breastfeeding.
Null hypothesis
Single and married men do not significantly differ in their attitude towards
exclusive breastfeeding.
Let the analysis be done straightaway and very quickly too.
Switch on the computer
Retrieve the data (Chapter 2 Table 1 Data.sav) that have since been
entered and securely saved
225
Analyze, using Compare Means.
Now that you have activated the SPSS Data Editor containing the data
file called Chapter 2 Table 1 Data.sav, the analysis proper is what has to
be done. Just in case you are yet to retrieve the file, move cursor to the
Menu Bar and select File Open Data and from the Open
Data dialog box, given in Figure 3.21, click the file name, Chapter 2 Table
1 Data.sav and click Open in the dialog box or press Enter on the
Keyboard to have the data file on your screen.
226
Figure 3.22 Analyze Compare Means Independent-Samples T Test
227
click the right-pointed arrow directly facing the Test Variable(s) box
at the right. Then, click on the independent variable (Marital status) to
highlight it in the left panel and click the right-pointed arrow directly
facing the box at the right that is called Grouping Variable. This transfers
the Marital status from the left box into the Grouping Variable panel at
the right as displayed in Figure 3.24.
Click Define Groups… to have its dialog box as shown in Figure 3.25
in which you provide or type the numerals with which each level or
category of the Grouping Variable (Marital status in this case) was
coded. Recall that while Single was coded 0, Married was coded 1 when
the data in question were named, labelled and provided values in Chapter 2.
228
Type 0, the numerical code for Single into the box beside Group 1; and
type 1, the numerical code for Married, in the box beside Group 2 as
shown in Figure 3.26.
229
Figure 3.27 Options dialog box
230
The first part of Output 3.3 that is known as Group Statistics has shown
that the Single (unmarried) men are 16 and they have a Mean of 49.5625,
Standard Deviation of 12.56433 and Standard Error Mean of 3.14108 on
Attitude Towards Exclusive Breastfeeding. Their counterparts, the married
men, are 16 with 60.0625 Mean, 11.87417 Standard Deviation and 2.96854
Standard Error Mean on the dependent variable (Attitude Towards
Exclusive Breastfeeding).
The second part of the Output 3.3, called Independent Samples Test,
has shown the Levene’s Test for Equality of Variances at the left-hand
side and the t Test for Equality of Means at the right-hand side. The
Levene’s Test for Equality of Variances has F of .165 and Sig. of .688,
meaning that the variances for the single and married men do not differ
significantly (F = .165, Sig. = .688, p < .05). Therefore, the t Test results to
be used are those for Equal variances assumed. For the Equal variances
assumed, the men’s attitude towards exclusive breastfeeding has a
computed t-value of -2.429, Degrees of Freedom of 30, Sig. (2-tailed) of
.021, Mean Difference of -10.50000, Standard Error of the Difference of
4.32188; and the Lower and Upper 95% Confidence Interval of the
Difference are -19.32646 and -1.67354, respectively.
Always note in accordance with Kpolovie (2018; 2011) that whenever
the statistical test for rejection or non-rejection of a null hypothesis is done
with the use of IBM SPSS Statistics, tenability of the null hypothesis is
determined by comparing the probability (p) value or significance level
(Sig.) in the output directly with the chosen alpha in the investigation. If the
p value is equal to or less than the chosen alpha, reject the null hypothesis.
Conversely, if the p value is greater than the chosen alpha, retain or do not
reject the null hypothesis. There is absolutely no reason to compare the
calculated statistical ratio arrived at via IBM SPSS Statistics with the
critical value of the statistical test as done when manual calculation is used
to test tenability of the null hypothesis.
Therefore, in this current analysis done that the p value (Sig.) of .021 for
2-tailed test is less than the chosen alpha of .05, the null hypothesis in the
example under consideration is rejected. That is, the null hypothesis that
“single and married men do not significantly differ in their attitude towards
exclusive breastfeeding” is rejected in favour of married men. With the
rejection of the null hypothesis, the alternate hypothesis that “there is a
statistically significant difference in the attitude of single and married men
231
towards exclusive breastfeeding” is sustained. The difference favours
married men who have significantly higher mean in their attitude towards
exclusive breastfeeding than the single men. The results of this analysis can
also be reported that the mean of single men’s attitude towards exclusive
breastfeeding that is equal to 49.5625 is significantly lower than the mean
of the married men towards exclusive breastfeeding that is equal to
60.0625. The results are aptly written technically that the difference in the
two means is statistically significant or simply that the null hypothesis is
rejected as t(30) = -2.429, p < .05, 2-tailed. In other words, the married
men have significantly better attitude towards exclusive breastfeeding than
the single or unmarried men.
It must as well be emphasized that whenever the Lower and Upper 95%
Confidence Interval of Difference both fall either above zero or below zero,
the pair of means under comparison statistically differ significantly. That is,
the mean difference between the two means (-10.50000) is statistically
significant. In the analysis that the Output is being interpreted here, both the
Lower and Upper 95% Confidence Interval of the Difference (-19.32646
and -1.67354, respectively) fall below zero. Therefore, the Mean Difference
of -10.50000 is statistically significant. On the contrary, whenever the
Lower 95% Confidence Interval of the Difference falls below zero and the
Upper 95% Confidence Interval of the Difference falls above zero, the pair
of means under comparison do not significantly differ. The mean difference
in the scenario that zero lies between the Lower and Upper 95% Confidence
Interval of the Difference is not significant and is totally a mere function of
chance.
Lastly, you must have observed in the Output 3.3 that apart from the t
ratio for Equal variances assumed, there is another t ratio of -2.429, df of
29.905, Sig. (2-tailed) of .021, -10.50000 Mean Difference, Standard Error
Difference of 4.32188, Lower value of -19.32764 and Upper -1.67236 for
95% Confidence Interval of the Difference for Equal variances not
assumed of the Levene’s Test for Equality of Variances. The rule for the
choice of which of the two t values (Equal variances assumed or Equal
variances not assumed) to use in testing tenability of a null hypothesis is
that whenever the significance (p value) of Levene’s Test is greater than .05
alpha, the t ratio for Equal variances assumed should be used. On the
contrary, whenever the significance of the Levene’s test is equal to or less
than .05 alpha, the t ratio and information under Equal variances not
232
assumed row should be used. That is, the values in the row for Equal
variances not assumed of the Levene’s Test for Equality of Variances
should be used only when the difference in the variances of the two
categories of the independent (grouping) variable is statistically significant.
With the Equation 3:3, the Effect Size, d, can easily be computed thus.
233
The d of .859 depicts a “large effect size”, as classified in Table 3.2, of
the independent variable (marital status) on the dependent variable (men’s
attitude towards exclusive breastfeeding). Men’s marital status largely
accounts for the variance in the men’s attitude towards exclusive
breastfeeding in favor of married men.
The results in Output 3.3 can therefore be summarized thus. The
married men’s attitude towards exclusive breastfeeding (M = 60.0625;
SD = 11.87417) significantly differs from the single (unmarried) men’s
attitude towards exclusive breastfeeding (M = 49.5625; SD = 12.56433);
and the Mean Difference (MD) = 10.50000: [t(30) = 2.429; p < .05 (two-
tailed); d = .859 (a large effect size) in favor of the married men.
Therefore, there is a significant difference in the attitude of men towards
exclusive breastfeeding as a function of marital status. Note that in
computing the t, if the lower mean (that of the unmarried men in this case)
is subtracted from the higher mean (that of the married men in the example
234
under consideration), there will not be any “–” sign in the computed t. In
fact, a “–” sign as part of a calculated t does not depict negativity, it only
shows that of the two groups under comparison, the one with the higher
mean was subtracted from the one with the lower mean. Thus, a minus sign
or no minus sign in a t value is only indicative of the side that the mean
difference favors. The married men have overwhelmingly better attitude
towards exclusive breastfeeding than the unmarried men.
Since the sample size of the married men (16) is equal to that of their
counterpart unmarried men (16), the Effect Size, d, can very easily be
computed with Equation 3:4. I have pointed out earlier that the Equation 3:4
states that the Cohen’s d as a measure of Effect Size for Independent-
Samples t Test is the mean difference between the two groups divided by
the square root of the average of the variances of the two groups.
Illustratively therefore, the Effect Size of the married and single men’s
attitude towards exclusive breastfeeding is as follows.
Where M2 and M1 are the respective means for the married men and
unmarried men; S22 and S12 are the variances (squares of the standard
deviations) for the two respective groups. Note that any of the two groups
(e.g., M2 and M1) could consistently come first in the equation. The d can
simply be computed as:
235
Apart from Equation 3:3 and Equation 3:4, irrespective of whether the
two groups are equal in sample size, Cohen’s d as a measure of Effect Size
for Independent-Samples t Test can be computed directly with Equation 3:5.
With Equation 3:5, the Effect Size for Independent-Samples t Test is:
Where N2 is the sample size of the group 2; N1 is the sample size of the
group 1; and t is the computed Independent-Samples t Test. In the current
example that the N2 is 16; N1 is 16; and t is 2.429, the d can be got as:
236
The d of .859 has already been interpreted as a “large effect size” of the
independent variable or the grouping variable (men’s marital status) on the
dependent variable (attitude towards exclusive breastfeeding). It must be
noted that in computing the d, I consistently subtracted the lower mean
from the higher mean to illustrate that in calculation of t and d, either of the
two means can consistently be subtracted from the other. When the lower
mean is subtracted from the higher mean, the answer will not have a “–” as
a prefix. But if the higher mean is subtracted consistently from the lower
mean, there will be a “–” sign in the answer to depict the side that the mean
difference favours.
Now that the interpretation of the output has been done, adhere to the
remaining routine protocols in data analysis with SPSS Statistics to
eventually shut down your system.
View the output
237
Shut down the system
In case there is any unit of these protocols that you are yet to master,
refer to previous sections and rehearse.
This Syntax as I have emphasized, is the same as the one given earlier
when Independent-Samples t Test was done on students’ Basic Science
Concepts with regard to Gender. The only exception is the names and
numerical codes for the variables as comparatively tabulated here.
Table 3.3 Syntax 3.1 and Syntax 3.2 compared
238
In Syntax 3.1, the independent, grouping or categorical variable is
Gender and was coded 1 and 2, while in Syntax 3.2, the independent,
grouping or categorical variable is Maritalstatus and was coded 0 and 1. In
Syntax 3.1, BSCS (Basic Science Concepts) is the dependent variable;
while in Syntax 3.2, the dependent variable is ATEB (Attitude Towards
Exclusive Breastfeeding). So, for any Independent-Samples t Test that you
will have to do with SPSS Statistics Syntax, simply replace the names and
codes of the variables in these examples with the names and codes of the
variables in your study when writing the SPSS Syntax.
Proceed with the analysis by observing the SPSS Statistics execution
conventions thus.
Retrieve the data (Chapter 2 Table 1 Data.sav) as they have since been
entered and securely saved.
Now that you have retrieved the SPSS Data Editor containing the data
file called Chapter 2 Table 1 Data.sav, the analysis proper is what has to
be done. But if you are yet to retrieve the file and is not very sure of how to
go about it; move cursor to the Toolbar and click the Open data document
icon and the Open Data dialog box appears as shown Figure 3.28.
239
Figure 3.28 Open Data dialog box
Look for and click the file name, Chapter 2 Table 1 Data.sav. That
click got the name of the saved dataset highlighted. Next, click Open in the
dialog box or press Enter on the Keyboard to have the document that was
saved as Chapter 2 Table 1 Data.sav opened on your screen.
With the data file open on the screen, move cursor to and click Window
Go to Designated Syntax Window. That is, click Window in the
Menu Bar to have the Window dropdown dialog box and in it, click Go to
Designated Syntax Window for the IBM SPSS Statistics Syntax Editor
to immediately appear as shown in Figure 3.29. The IBM SPSS Statistics
Syntax Editor can also be got by clicking File New Syntax.
That is, click File and in the File dropdown menu, select New; and in the
dropdown of the New, click Syntax.
240
Figure 3.29 IBM SPSS Statistics Syntax Editor
Most carefully type in the IBM SPSS Statistics Syntax that is given in
Syntax 3.2. When entered, the SPSS Syntax Editor will look like the
screenshot that is shown in Figure 3.30.
Figure 3.30 SPSS Syntax Editor with the Syntax for Independent-Samples t Test
Click Run and click All (see Figure 3.31) to have the results displayed.
Alternatively, select and highlight the entered Syntax and click the Run
241
Output 3.4 Chapter 3 Output 4.spv
This Output 3.4 is exactly the same as Output 3.3. The same
interpretations of Output 3.3 are most appropriately applicable with the
Output 3.4.
View the entire output by scrolling left, right, up and down with the left-
pointed and right-pointed arrows at the bottom edge of the screen, and with
the up-pointed and down-pointed arrows at the right edge of the IBM SPSS
Statistics Viewer.
Save the output. Move cursor to the Toolbar and click the Save this
document icon in the Toolbar for the Save Output As… dialog box
to appear. In it, type the name (Chapter 3 Output 4.spv in this case) with
which you want the file to be saved in the box beside File name and click
Save or press Enter as shown in Figure 3.32.
242
Figure 3.32 Save Output As… Chapter 3 Output 4.spv
Also, get to the Syntax Editor with the entered Syntax and click the
Save this document icon in the Toolbar to have its dialog box. Type
in the name, Chapter 3 Syntax 2.sps, with which you want to save the
Syntax (see Figure 3.33). Then, press Enter on the Keyboard or click Save
in the dialog box to have the Syntax saved.
243
Figure 3.33 Save Syntax As… Chapter 3 Syntax 2.sps
Print the output by clicking the Print icon in the Toolbar when the
SPSS Viewer with the Output 3.4 is activated (see Figure 3.4), and click
OK to get one copy of the Output 3.4 printed. Repeat the operation to get
the required number of copies printed.
244
Figure 3.34 Print All visible output menu
Exit IBM SPSS Statistics by closing all the opened SPSS windows. On
getting to the last Data Editor that is opened, clicking Close produces a
dialog box as exhibited in Figure 3.35 to prompt you to confirm if really
you want to exit IBM SPSS Statistics. In it, click Yes.
245
Chapter 4
ONE-SAMPLE t TEST EXECUTION
Overview
In research scenarios that the population mean of the variable of interest is
known, the investigator is required to draw a single sample and compare it
with the parameter for possible establishment of significant difference, if it
does exist, the one-sample t Test is used in the data analysis. Execution of
such parametric inferential comparisons with SPSS Statistics is
magnificently demonstrated with screenshots from data entry to results
interpretation for the user to absolutely acquire expert skills in performing
One-sample t Test. Life expectancy, and private dedicated gifted schools’
students’ IQ are used for the demonstration.
246
Taking Life Expectancy that the World average is known as a dependent
variable, it is possible to draw one sample from the population and compare
the sample mean with the world population mean. The World average,
termed Test Value in One-Sample t Test, of Life Expectancy is 72.24
(Kpolovie, Oshodi & Iwuchukwu, 2016). Life Expectancy is a statistical or
theoretical measure of the length of time or years an average human in a
given country and continent is supposed to live on earth. Life Expectancy is
an estimate of the average age at which individuals in a particular country
will be when they die. It is the average length of years that the citizens of a
country live before their death. To illustrate computation of One-Sample t
Test, the Life Expectancy of 44 randomly sampled European countries will
be compared with the known world average of 72.24 to ascertain whether
humans are expected to live significantly longer or shorter in Europe than
the world average. In other words, the data collected are meant for finding
out how long humans are expected to live before their death in Europe and
in the World. The data are presented in Table 4.1.
247
Research question, Alternate hypothesis and Null
hypothesis
The research question asked and the hypothesis postulated for the
investigation are as follows.
Research question:
248
What is the difference between life expectancy in Europe and in the
world?
Alternate hypothesis:
Life expectancy in Europe differs significantly from that of the world.
Null hypothesis:
A significant difference does not exist between life expectancy in
Europe and the life expectancy in the world.
Step 2. Launch IBM SPSS Statistics by clicking its icon . Wait for it to
fully get ready and when it is done, close the superimposed dialog box on
the Data Editor to be left exactly with the IBM SPSS Statistics Data
Editor alone on the screen.
Step 3. Enter the data. Begin with the topmost and leftmost cell in the SPSS
Data Editor by clicking on that particular cell. Type each value and press
the Enter key on the Keyboard. Note that each value represents the life
expectancy of a particular country in Europe, and the countries are
represented with the serial numbers in the Data Editor. On completion of
the data entry in the Data View, click Variable View. In the Name column
of the Variable View, type the variable name, LExEurope, without space.
In the column for Label, type the exact label for the variable as Life
Expectancy in Europe. Ensure that Scale is written for the row under the
column, Measure, as shown in Figure 4.1.
249
Once you click Data View at the extreme bottom left of the IBM SPSS
Data Editor, it takes you back to the Data View with the variable properly
named and with all the entered data. So, click Data View to confirm what
happens as presented in Figure 4.2.
Figure 4.2 Data Editor with the data and named variable
250
Figure 4.2 Continued
Step 4. Save the data. Move cursor to the Toolbar and click the Save this
document icon . That click instantly produces the Save Data As dialog
box. In the File name field within the Save Data As dialog box, type a
suitable name, Chapter 4 Table 1 Data.sav, in this example and click
251
Save. Alternatively, move cursor to the Menu Bar and click File
Save As… (illustrated in Figure 4.3) for the Save Data As dialog box to
appear for you to type in a name for the document. So, name the file as
Chapter 4 Table 1 Data.sav as exemplified in Figure 4.4 and click Save or
press Enter to have the dataset saved properly for easy retrieval as at when
needed.
252
Figure 4.4 Save Data As dialog box with the file name
Step 5. Analyze the data. It is at this step that the entire statistical operations
for execution of the One-Sample t Test are performed.
Click Analyze
Click Compare Means (see Figure 4.5)
Click One-Sample t Test. This click produces the One-Sample T
Test menu that is exhibited in Figure 4.6.
253
Figure 4.5 Analyze Compare Means One-Sample
254
Figure 4.7 One-Sample T Test with the variable moved right
In the box beside Test Value, type in the population mean, that is the
world average of Life Expectancy which is 72.24 as shown in Figure
4.8.
Figure 4.8 One-Sample T Test menu with the Test Value entered
255
INTERPRETATION OF OUTPUT 4.1: LIFE
EXPECTANCY IN EUROPE
The first part of the One-Sample t Test output, termed One-Sample
Statistics, has shown the sample size (N) to be 44, the Mean of Life
Expectancy in Europe to be 79.3214, Standard Deviation to be 3.51484 and
the Standard Error of Mean to be .52988. These descriptive statistics could
serve as the answer to the research question by stating the sample mean
against the population mean.
The second part of Output 4.1, called One-Sample Test, has provided
the Test Value of 72.24 (which is the World mean), the computed t-ratio to
be 13.364, degrees of freedom (df) to be 43 (number of cases minus 1), Sig.
(2-tailed) to be .000 that is interpreted or written as less than .0005, and
Mean Difference of 7.08136. In addition, the Lower and Upper 95%
Confidence Interval of the Difference are respectively 6.0128 and 8.1500.
The computed t is actually arrived at by dividing the Mean Difference
(7.08136) by the Standard Error of Means (.52988). That is, t = Mean
256
Difference ÷ Standard Error of Mean (t = 7.08136 ÷ .52988 = 13.364).
The p-value (Sig. 2-tailed) of less than .0005 is smaller than the chosen
alpha of .01. Therefore, the null hypothesis that “a significant difference
does not exist between Life Expectancy in Europe and the Life Expectancy
in the World” is rejected in favour of European countries. That is, European
countries have Life Expectancy mean (79.3214) that is significantly greater
than the average of Life Expectancy in the World (72.24). With the rejection
of the null hypothesis, the alternate hypothesis that “Life Expectancy in
Europe differs significantly from that of the World” is sustained as t(43) =
13.364, p < .01.
It must finally be emphasized that whenever the Lower and Upper 95%
Confidence Interval of the Difference both fall either above zero or below
zero, the pair of means under comparison statistically differ significantly.
That is, the mean difference between the two means is significant. In the
analysis that the results (Output 4.1) is being interpreted here, both the
Lower and Upper 95% Confidence Interval of the Difference (6.0128
and 8.1500, respectively) fall above zero. Therefore, the Mean Difference
of 7.08136 is statistically significant. Substantially high Effect Size is
expected as the mean difference is significant even at .0005 alpha.
On the contrary, whenever the Lower 95% Confidence Interval of the
Difference falls below zero and the Upper 95% Confidence Interval of
the Difference falls above zero, the pair of means under comparison do not
significantly vary. The mean difference in the scenario that zero lies
between the Lower and Upper 95% Confidence Interval of the
Difference is not significant and is interpreted as a mere function of chance
as its Effect Size is bound to be “trivial”.
257
“Trivial”, “Small”, “Medium”, or “Large” Effect Size. Recall the
classification of Effect Size that I have given in earlier in Table 3.2 of
Chapter 3 to be as follows. These classifications correspond with the
recommendations by Cohen (1988).
d AS EFFECT SIZE DESCRIPTIVE
IN t Test CLASSIFICATION d
0.000 to 0.199 Trivial Effect Size
0.200 to 0.499 Small Effect Size
0.500 to 0.799 Medium Effect Size
0.800 and above Large Effect Size
Cohen’s d as the measure of Effect Size for the One-Sample t Test can,
and should be computed as the Mean Difference between the population
mean and the sample mean divided by the sample Standard Deviation. In
the current example, while the Life Expectancy Mean Difference is
7.08136, the sample Standard Deviation is 3.51484. Therefore, the Effect
Size, d, is as given in Equation 4:1:
A d of 2.015 is, indeed, a very “Large Effect Size”. Thus, Europe has a
Life Expectancy (79.3214) that is significantly higher than the world’s Life
Expectancy (72.24); and the Mean Difference is equal to a vary “Large
Effect Size” (2.015). The results can therefore be fully summarized as Life
258
Expectancy in Europe (M = 79.3214 and SD = 3.51484) is significantly
better than the Life Expectancy in the World (M = 72.24): [t(43) =
13.364; p < .0005; d = 2.015 (a “Large Effect Size”)].
Apart from the Equation 4:1, d as the measure of Effect Size in One-
Sample t Test can be computed directly from the calculated One-Sample t
Test ratio as the t value multiplied by the square root of 1 divided by the
sample size. In equation form, computing d directly from the t value is as
given in Equation 4:2.
Either of the two equations (Equation 4:2 and Equation 4:1) can be used
for determination of the Effect Size whenever One-Sample t Test is
executed in a research. Application of the two equations is even better
because each can be used for confirmation of the other as they are bound,
by right, to produce the same result just as each of them in this example has
demonstrated that the Effect Size of the difference between the Life
Expectancy in Europe and in the World is 2.015 that is a “Large Effect
Size” in favour of Europe.
With the completion of the interpretation of the Output 4.1, it is time to
proceed with the remaining routine protocols of data analysis with IBM
259
SPSS Statistics and eventually shut down the system.
Step 6. View the output by scrolling right. left, up and down with the arrows
at the bottom edge and right edge of the screen.
Step 7. Save the output. While on the IBM SPSS Statistics Viewer
window, click the Save this document icon in the Toolbar to have the
Save Output As dialog box as exemplified in Figure 4.9. Type a suitable
name (Chapter 4 Output 1.spv) in the box beside File name as shown in
Figure 4.10, and click Save or press the Enter key on the Keyboard.
260
Figure 4.10 Save Output As with the document name
Step 8. Print the output and input (data) by clicking the Print icon in
the Toolbar and clicking OK to get one copy of All visible output printed.
Repeat the action until the desired number of copies are printed. To print the
input or data, activate the IBM SPSS Statistics Data Editor with the data
file (Chapter 4 Table 1 data.sav) on the screen, click the Print tool
in the Toolbar and click OK to get a copy of the data file printed.
Step 9. Exit SPSS Statistics by simply closing every SPSS window that is
open on the computer screen.
Step 10. Shut down the system by clicking Start Power Shut
down.
261
SPSS SYNTAX EXECUTION OF ONE-SAMPLE t
TEST
Use of IBM SPSS Statistics Syntax is a very easy and fast method for
execution of One-Sample t Test as illustrated in this section. Correct typing
in of the statistical Commands in the SPSS Syntax Editor is everything
required. The Syntax for performing One-Sample t Test is as given in
Syntax 4.1.
Just follow the simple routine steps for data analysis with SPSS.
Switch on the computer
Retrieve the data since the entry has already been done.
Now that you have activated the SPSS Data Editor containing the data
file called Chapter 4 Table 1 Data.sav, the analysis proper is what has to
be done. In case you have not successfully retrieved the file, move cursor to
the Menu Bar and click File Open Data (see Figure 4.11) to
have the Open Data dialog box as given in Figure 4.12.
262
Figure 4.11 File Open Data
263
Click the file name, Chapter 4 Table 1 Data.sav, and click Open in the
dialog box or press Enter on the Keyboard to have the data file on your
screen as earlier shown in Figure 4.2.
Move cursor to the Menu Bar and click Window Go to
Designated Syntax Window that is demonstrated in Figure 4.13.
Once you click the Go to Designated Syntax Window, SPSS takes you
there instantly as shown in Figure 4.14.
264
Figure 4.15 IBM SPSS Statistics Syntax Editor with the Syntax entered
Click Run ALL (see Figure 4.16), or highlight the entered Syntax
and click the Run icon in the Toolbar. With this, IBM SPSS Statistics
instantly completes the analysis and displays the results in the SPSS
Viewer window as exemplified in Output 4.2.
265
INTERPRETATION OF OUTPUT 4.2
The same interpretation for Output 4.1 when the SPSS Dialog Boxes Point
and Click method was used for analysing the European life expectancy data
is applicable with the Output 4.2 when SPSS Syntax method was adopted
for execution of the t Test.
Follow the remaining routine protocols to eventually shut down your
system.
View output
Save the output as Chapter 4 Output 2.spv and the Syntax as Chapter 4
Syntax 1.sps
Print output
266
Exit IBM SPSS Statistics (see Figure 4.17)
Since you have completed the analysis that you wanted to perform, click
Yes to exit SPSS. If there is any unit of these protocols that you are yet to
master, refer to previous sections and rehearse.
267
States average in the WJ III consists of standard score range of 90 to 110
that is equivalent to 25 to 75 percentile rank. The psychometrician
randomly drew a sample of 66 students from the three private Dedicated
Gifted Schools, namely:
i. Hampshire Country School
ii. North Broward Preparatory School
iii. New Mexico Military Institute
The psychometrician administered both the General Intellectual
Ability (GIA) and Brief Intellectual Ability (BIA) Scales of the WJ III to
the sampled students. These two Scales contain all the nine Sub-Scales of
the WJ III:
1. Comprehension Knowledge
2. Long-Term Retrieval
3. Visual-Spatial Thinking
4. Auditory Processing
5. Fluid Reasoning
6. Processing Speed
7. Short-Term Memory
8. Quantitative Knowledge
9. Reading Writing
A sample of 66 students from the three private Dedicated Gifted Schools
was suitably large enough for the study. For instance, Hampshire Country
School in the 2018/2019 session runs 3-10 Grades with 21 students and an
average class size of 4 students. The North Broward Preparatory School
programme range from Prekindergarten to 12 with student population of
1445 and an average class size of 15 students. Of the 1445 students, those
aged 9 to 12 are relatively few. New Mexico Military Institute runs Grade 9
to 12 with a student population of 500 and an average class size of 14
students. Each of these three boarding schools has over the years proven to
live above board, honing leadership skills and character through exceptional
education, training, experience and an uncommon academic excellence. It is
one of the best opportunities that a child could get to be a student in any of
these three schools and discover his/her talents and special aptitudes.
With the use of Intelligence Quotient tests, gifted children are not only
determined, but are classified into five categories for purposes of suitable
special accelerated educational programmes and other unique support
268
(Davis, Rimm & Siegle, 2011; Kpolovie, 2017; 2015; 2011). The five
classification of giftedness on the basis of demonstrated IQ are the:
Profoundly gifted with IQ of 180 and above
Exceptionally gifted with IQ of 160 – 179
Highly gifted with IQ of 145 – 159
Moderately gifted with IQ of 130 – 144
Mildly gifted with IQ of 115 – 129.
It is expected that if students in the privately Dedicated Gifted Schools
are indeed gifted, then their IQ on the average will at least be 15 points
higher than the population IQ mean of 100. That is, the private dedicated
Gifted School students will on the average be having IQ that range from
115 and above as measured with the Woodcock-Johnson Test of Cognitive
Abilities (WJ III).
On the whole, only about 32 million students in the United States have
IQ that qualifies them to be classified as gifted and are treated accordingly
with accelerated individualized educational curriculum (Alamieyeseigha
and Kpolovie, 2013). Approximately 68% of the US student population
have IQ that range from 85 to 115 and fall within ±1z on a normal curve;
and they are the students who are actually addressed by the standard school
age/grade-based curriculum. Students whose IQ do not fall within the -1z
and +1z require completely different curriculum (Kpolovie, 2012; 2015;
2017; Davis, Rimm & Siegle, 2011).
Between +1z and +2z of the normal curve has IQ that range from 115 to
130 that is scored by approximately 13.59% of the student population.
These students are gifted and demand a special curriculum that is greatly
modified accordingly to meet their learning needs. They are in the main,
accorded such curriculum. Between +2z and +3z of the normal curve has IQ
range of 130 to 145 that only 2.14% of the US student population have.
These are exceptionally gifted students that only very unique individualized
accelerated curriculum can suitably take care of. Between +3z and +4z of
the normal curve has .13% of the student population, denoting 13 students
in every 10,000, who have IQ that range from 145 to 160. These students
are so exceedingly gifted that only highly exceptional individualized
curriculum could meet their unique needs. At the very extreme above the
mean, +4z and beyond, has .003% of the student population who have IQ
that is greater than 160 to depict their profound giftedness. Only one in
every 30,000 students in the population could possess such extreme
269
giftedness. Accordingly, only exceptionally extreme individualized
accelerated curriculum could meet the needs of such profoundly gifted
students. In the US, each of these categories of students are actually
provided with the unique curriculum that best matches them; unlike the
Nigerian situation that every student is forced to undergo the same
curriculum to become and remain nothing more than an average person
(Alamieyeseigha & Kpolovie, 2013; Kpolovie, 2012).
The normal curve of IQ is indeed normally distributed symmetrically
(Herrnstein & Murray, 1994; Kpolovie, 2017; 2015; Hunt, 2011) and so, the
same 13.59% of the students have IQ that range from 70 to 85 and fall
within -1z and -2z. These students also need a different curriculum that will
be most suitable for them. Two standard deviations below the mean, within
-2z and -3z of the normal curve also has just 2.14% of the student
population with IQ scores that range from 55 to 70; and these exceptional
students need their own customized individualized curriculum that best
addresses each of them. In like manner, between -3z and -4z of the normal
curve has .13% of students with IQ that is less than 55, and they require a
most special individualized curriculum to address their peculiar
exceptionality. Also, there is .003% of the student population with IQ that
falls lower than -4z below the mean. Exceptionally extreme individualized
curriculum is necessarily needed for such students to bring out the best in
them and make them more useful and less dependent.
Funding of education must be structured to adequately accommodate the
unique needs of the different types of students. This necessitates
investigation of the extent to which students in the private Dedicated Gifted
Schools differ from the United States average students with respect to their
IQ as measured with the Woodcock-Johnson Test of Cognitive Abilities
(WJ III). The scores of the sampled students from the Dedicated Gifted
Schools on the Woodcock-Johnson Test of Cognitive Abilities (WJ III) are
given in Table 4.2 and are used to illustrate how the One-Sample t Test
analysis is performed with the use of IBM SPSS Statistics.
270
3 113 36 134
4 145 37 142
5 128 38 127
6 131 39 115
7 122 40 112
8 136 41 148
9 113 42 151
10 146 43 109
11 125 44 161
12 127 45 122
13 148 46 155
14 109 47 111
15 156 48 142
16 124 49 130
17 132 50 138
18 135 51 130
19 108 52 142
20 151 53 161
21 120 54 123
22 125 55 134
23 137 56 140
24 138 57 141
25 110 58 123
26 144 59 144
27 137 60 153
28 150 61 142
271
29 148 62 133
30 120 63 144
31 126 64 133
32 116 65 131
33 117 66 122
Research question
What is the difference between private Dedicated Gifted Schools Students’
WJ III mean IQ and the United States students’ WJ III mean IQ?
Alternate hypothesis
Students in the private Dedicated Gifted Schools differ significantly from
the general United States students in their mean IQ scores on the
Woodcock-Johnson Test of Cognitive Abilities (WJ III).
Null hypothesis
Students in the private Dedicated Gifted Schools do not differ significantly
from the general United States students in their mean IQ scores on the
Woodcock-Johnson Test of Cognitive Abilities (WJ III).
272
Enter the data very carefully as done in the earlier example, name and
label the variable as shown in Figure 4.18
273
Figure 4.18 Chapter 4 Table 2 Data.sav
274
Figure 4.18 Continued
275
Figure 4.18 Continued
276
Figure 4.19 Save Data As… Chapter 4 Data 2.sav
277
Figure 4.20 Analyze Compare Means One-Sample T Test
278
Figure 4.21 One-Sample T Test menu ready for output
For IBM SPSS Statistics to display the results of the analysis in the
SPSS Viewer window, click OK. The results are as exemplified in Output
4.3.
279
INTERPRETATION OF OUTPUT 4.3: STUDENTS’
WJ III IQ
It is clear from the One-Sample Statistics table in Output 4.3 that the IQ of
the 66 students from Dedicated Gifted Schools on the Woodcock-Johnson
Test of Cognitive Abilities, 3rd Revision (WJ III IQ), has a Mean of
132.5909, Standard Deviation of 13.97688 and Standard Error of Mean of
1.72043.
It can be discerned from the One-Sample Test table of the Output 4.3
that the Test Value is 100 and that the t is 18.943 with 65 degrees of
freedom (df), .000 p-value (Sig. 2-tailed) that is written as less than .0005,
a 32.59091 Mean Difference, plus 29.1550 Lower and 36.0269 Upper
95% Confidence Interval of the Difference. Since the p-value (Sig. [2-
tailed]) of .0005, is smaller than the chosen .05 and even of .001 level of
significance, the null hypothesis that “students in the Dedicated Gifted
Schools do not differ significantly from the general United States students
in their mean IQ scores on the Woodcock-Johnson Test of Cognitive
Abilities (WJ III)” is rejected. The mean difference between the Dedicated
280
Gifted Schools students and the United States population mean on the WJ
III IQ is 32.59091 that is statistically significant. In summary, with regards
to the Woodcock-Johnson Test of Cognitive Abilities, the IQ of students
from the Dedicated Gifted Schools (M = 132.5909, SD = 13.97688, SEM =
1.72043) is significantly higher than the United States student population
mean of 100 because [t(65) = 18.943, p < .0005].
It must finally be emphasized that whenever the Lower and Upper 95%
Confidence Interval of the Difference both fall either above zero or below
zero, the pair of means under comparison statistically differ significantly.
That is, the mean difference (32.59091) between the population mean (i.e.,
the Test Value of 100) and the Sample mean of 132.5909 is significant. In
the analysis that the output is being interpreted here, both the Lower and
Upper 95% Confidence Interval of the Difference (29.1550 and 36.0269,
respectively) fall far above zero. Therefore, the Mean Difference of
32.59091 is statistically significant.
On the contrary, whenever the Lower 95% Confidence Interval of the
Difference falls below zero and the Upper 95% Confidence Interval of
the Difference falls above zero, the pair of means under comparison do not
significantly vary. The mean difference in the scenario that zero lies
between the Lower and Upper 95% Confidence Interval of the
Difference is not significant and is totally a mere function of chance.
But the current output under interpretation in which both the Lower and
Upper 95% Confidence Interval of the Difference fall totally above zero has
demonstrated a statistically significant difference between the sample mean
and the population mean in favour of the sample (the private Dedicated
Gifted Schools students). The private Dedicated Gifted Schools students, on
the average, have IQ that is more than 32 IQ points higher than that of the
US population IQ mean of 100. So, the students in private Dedicated Gifted
Schools are indeed gifted.
281
Recall that Effect Size, measured with Cohen’s d, is classified as:
Less than .200 is interpreted or described as “Trivial Effect
Size”;
Between .200 and .499 is described as “Small Effect Size”;
Between .500 and .799 is classified as “Medium Effect Size”; and
Up to .800 and above is interpreted as “Large Effect Size”.
These classifications corroborate the recommendations by Cohen
(1988).
Cohen’s d as the measure of Effect Size for the One-Sample t Test is
computed as the Mean Difference between the population mean and the
sample mean divided by the estimated population Standard Deviation
which is, indeed, the sample Standard Deviation in a One-Sample t Test.
In this example, while the Mean Difference is 32.59091, the sample
Standard Deviation is 13.97688. Therefore, the Effect Size, d, is as given
in Equation 4:1:
282
Where t is the calculated value of the One-Sample t Test, and N is the
size of the sample. In the current illustration, the t value is 18.943 and the
sample size, N, is 66. Replacing the symbols in the Equation 4:2 with their
actual values, the Cohen’s d as the measure of Effect Size (ES) is:
283
Shut down the system
That is, complete the remaining steps thus:
View the output by scrolling right, left, up and down with the arrows at
the bottom edge and right edge of the screen.
Save the output. While on the IBM SPSS Statistics Viewer window,
click the Save this document icon in the Toolbar to have the Save
Output As… dialog box as exhibited in Figure 4.22 and type a suitable
name (Chapter 4 Output 3.spv) in the box beside File name as shown in
Figure 4.23, and press the Enter key on the Keyboard or click Save.
284
Figure 4.23 Save Output As with the document name
Print the output and the data inputted by clicking the Print icon in
the Toolbar and clicking OK to get a copy of All visible output printed. To
print the input or data, activate the IBM SPSS Statistics Data Editor with
the dataset (Chapter 4 Table 2 Data.sav) on the screen. Then, click the
Print tool in the Toolbar and click OK to get the data file printed.
Exit SPSS Statistics by simply closing every SPSS document that is
open on the computer screen.
Shut down the system by clicking Start Power Shut
down. That is, move cursor to and click Start that is located at one of the
corners on the screen (left bottom corner usually as default), click Power,
and click Shut down.
285
ONE-SAMPLE t TEST WITH SYNTAX FOR
IQ_WJIII DIFFERENCE
The IBM SPSS Statistics Syntax for execution of the current analysis is
exactly like the one in Syntax 4.1 with just a little but critically important
modification of replacing the Variable Name and the Test Value as
indicated in Syntax 4.2. Whenever you are to perform One-Sample t-test
analysis with IBM SPSS Statistics Syntax method, it is this same Syntax
that should be used with the little but very crucial modification of replacing
the Test Value and the Variable Name here with the exact Test Value (the
population mean) and the exact name of the variable that your data are on.
Needless further narrative, just get the analysis done with the provided
Syntax 4.2. It is the execution of the analysis that truly matters.
Switch on the computer
Retrieve the data since the entry has already been made.
Retrieve the dataset earlier saved as Chapter 4 Table 2 Data.sav. The
dataset retrieval demands that when you are on the IBM SPSS Statistics
Data Editor, move the cursor to the Toolbar and click the Open this
document icon and its dialog box appears as illustrated in Figure 4.24.
In the Open Data menu, look for the name of the file you want to retrieve
(Chapter 4 Table 2 Data.sav in this case), click the name of the file and it
becomes highlighted. Then, press Enter on the Keyboard or click Open.
286
On clicking Open or pressing Enter, the document opens as shown earlier
in Figure 4.18.
287
clicking Window Go to Designated Syntax Window, the SPSS
Statistics Syntax Editor can also easily be made to appear by clicking File
and in the File dropdown menu, selecting New, and finally clicking Syntax
in the New dropdown menu.
Most carefully enter the Syntax provided in Syntax 4.2 into the SPSS
Syntax Editor and it will look as illustrated in Figure 4.27.
Finally, highlight the entered Syntax and click the Run icon in the
Toolbar. This action produces the results. Otherwise, do not highlight the
entered Syntax, but just click Run on the Menu Bar and in the Run
dropdown menu, click All for IBM SPSS Statistics to complete the analysis
in a split second and produce the results in the IBM SPSS Statistics
Viewer as displayed in Output 4.4.
288
INTERPRETATION OF OUTPUT 4.4
The same interpretation advanced for Output 4.3 is applicable here as the
Output 4.4 and Output 4.3 are exactly the same. The same dataset (Chapter
4 Table 2 Data.sav) was analysed to produce the equivalent results that
attract the identical interpretation. The only difference between the two
Outputs (4.4 and 4.3) is the process adopted in executing the analysis.
While IBM SPSS Statistics Syntax method was used to arrive at Output
4.4, IBM SPSS Statistics Dialog Boxes Point and Click approach was
employed to arrive at Output 4.3.
Though it is necessary to learn and master both techniques for the
analysis, the decision on which of the two methods to apply in performing a
particular One-Sample t Test analysis depends on you, the user. No matter
the method that you choose to use, it is always better to know more than
one technique for execution of each statistical analysis. If ordinary mouse
will usually have more than one hole, why won’t you master more than one
approach for analysing each statistical test?
View the output by scrolling up, down, right and left, using the requisite
arrows on the screen for the purpose.
289
Save the Output by clicking the Save this document icon on the
Toolbar as shown in Figure 4.28.
Type in a suitable file name, Chapter 4 Output 4.spv, and click Save or
press the Enter key on the Keyboard.
You need to also save the Syntax. Activate the Syntax Editor
containing the Syntax with which the analysis was done. Click the Save
this document icon on the Toolbar to get its menu as given in Figure
4.29. Type in the Syntax name (Chapter 4 Syntax 2.sps) and press Enter
or click Save.
290
Figure 4.29 Save Syntax As Chapter 4 Syntax 2
Print the Output by clicking the Print icon in the Toolbar and
clicking OK as can be seen in Figure 4.30 for one copy of the output to be
printed. The printing operation can be repeated until the desired copies of
the output are all printed. Alternatively, the desired number of copies of the
output can be printed at once by using the up-pointed arrow beside Number
of copies: to accordingly increase the number of copies before
clicking the OK.
291
Figure 4.30 Print All visible output
Exit IBM SPSS Statistics by closing all opened windows, and opting to
indeed Exit SPSS Statistics.
Shut down the system by clicking Start Power Shut
down.
292
Chapter 5
PAIRED-SAMPLES t TEST
EXECUTION
Overview
Investigations that deal with the comparison of two means from repeated
measures, dependent, correlated or matched samples demand application of
Paired-samples t Test. In a single group pretest-posttest experiment, each
subject is observed twice. The means from the repeated measures can best
be subjected to the paired-samples t Test for accurate drawing of inference
on the effect of the experimental treatment to validly cover the population
from which the representative samples were drawn. This chapter empowers
the user on execution and interpretation of correlated-samples t Test with
both SPSS dialog boxes selection and syntax methods. Spouses’
educational attainment, and self-actualization at the points of admission into
and graduation from a 3-year degree program are used for illustration.
293
related individuals could be identical twins, fraternal twins, spouses
(husbands and their wives), mothers and their daughters, fathers and their
sons, first and second siblings, and chancellors and their vice-chancellors.
Chiefly, for comparison of two means from two samples that are naturally
related in some way, the Paired-Samples t Test must be used. There are
several investigations that could demand application of Paired-Samples t
Test. Two examples are provided in this chapter on how Paired-Samples t
Test is performed, using IBM SPSS Statistics dialog boxes point and click
method and IBM SPSS Statistics Syntax approach in each case.
294
Bachelor’s degree 100
Master’s degree 110
Professional degree 120
Doctoral degree 130
295
21 100 100
22 110 100
23 110 100
24 120 110
25 100 110
26 120 120
27 130 100
28 130 100
29 120 110
30 120 80
31 110 90
32 120 130
33 110 110
34 100 90
35 120 100
36 90 100
37 130 100
38 120 100
39 120 100
40 100 100
41 100 110
42 100 90
43 110 80
44 120 100
45 100 90
46 90 100
296
47 90 100
48 80 110
49 110 90
50 100 80
51 120 70
52 100 110
53 100 100
54 100 100
55 130 80
56 110 100
57 100 100
58 100 130
59 90 100
60 130 100
61 80 100
62 90 110
63 120 100
64 80 90
65 120 100
66 110 120
297
Husbands and their wives do not differ significantly in educational
attainment.
Step 3. Enter the data. Entering of data in IBM SPSS Statistics for
performing of Paired-Samples t Test is dramatically different from how
data are inputted for execution of Independent-Samples t Test that was
treated in Chapter 3. For Paired-Samples t Test, while the participants’
first scores on the dependent variable (Spouses’ Educational Attainment
in this instance) are entered in the first column of the SPSS Data View, the
participants’ second scores are entered in the second column of the SPSS
Data View in such a careful manner that the two scores of one natural pair
(i.e., husband and wife in this example) of participants must be on the
same row. Emphatically, each couple’s two scores must be on the same row
for that particular family, without any mistake or shift for anyone.
In Paired-Samples t Test, the pairs or two sets of score are not
differentiated with codes of 1 and 2 or 0 and 1 as was done in
Independent-Samples t Test. Rather, all the scores on one category of the
respondents must be entered under the first column that IBM SPSS
Statistics default describes as Var00001, and all the scores of the other
category of respondents must be inputted under the second column that
SPSS default regards as Var00002, ensuring unmistakably that every
respondent’s two scores constitute one and the same row.
Now, you can start entering the scores, one after the other and making
sure that after typing each score, the Enter key on the Keyboard is pressed
once. After entering the data accordingly, the relative positions (order and
sequence) of the data for husbands and wives will be as serially numbered
in Table 5.1. Then, click Variable View that will be looking like Figure 5.1.
298
Figure 5.1 Variable View with default names for the variables
Replace the Var00001 and Var00002 with their actual names, Husband
and Wife, respectively under the Name column. In the first and second
rows under Label column, type Husband and Wife, respectively. Under
Measure column, click the Unknown and select Scale for the first and
second rows. When done, the Variable View will look like the screenshot
in Figure 5.2.
Click Data View at the extreme bottom and extreme right to return to
the Data View when all the data have properly been entered, named and
labeled as shown in Figure 5.3.
299
Figure 5.3 Chapter 5 Table 1 Data.sav
300
Figure 5.3 Continued
301
Figure 5.3 Continued
Step 4. Save the data by moving cursor to the Toolbar and clicking the Save
this document icon for Save Data As dialog box to appear. In the Save
Data As dialog box, type a suitable unique name for the dataset in the File
name field. This dataset is accordingly named Chapter 5 Table 1 Data.sav
as can be seen in Figure 5.4.
302
Figure 5.4 Save Data As dialog box
303
Figure 5.5 Analyze Compare Means Paired-Samples T Test
304
With Husband highlighted in the left box, click the right-pointed arrow
to move the variable into the Paired Variables box at the right. Also
move the remaining variable, Wife, from the left box into the Paired
Variables box at the right by highlighting Wife and clicking the same right-
pointed arrow . When done, the Paired-Samples T Test dialog box
will look like Figure 5.7. Preferably, highlight both of the variables
(Husband and Wife) in the left box and click the right-pointed arrow
to transfer the two variables at once into the Paired Variables field at the
right to have the Figure 5.7.
Figure 5.7 Paired-Samples T Test dialog box ready for production of output
Finally, click OK. With this click, the results of the Paired-Samples t
Test on Spouses’ Educational Attainment (SEA) in California will appear in
the IBM SPSS Statistics Viewer as presented in Output 5.1.
305
INTERPRETATION OF OUTPUT 5.1: COUPLES’
EDUCATIONAL ATTAINMENT
The first part of the Output 5.1, Paired Sample Statistics, reveals the
descriptive statistics on Spouses’ Educational Attainment (SEA) for
Husband and Wife. For Husband, the Mean is 107.8788, N (number of
cases) is 66, Standard Deviation is 13.64553 and Standard Error of
Mean is 1.67965. For Wife, the Mean is 98.9394, N is 66, Standard
Deviation is 11.51972 and 1.41798 Standard Error of Mean. The
Husband’s mean tends to be higher than that of the Wife. Whether the
difference between these means is statistical and not by chance occurrence
shall be unveiled in the third part (Paired Samples Test) of the output.
The second table in the Output 5.1, Paired Samples Correlations,
shows that the Pearson correlation coefficient between the Spouses’
Educational Attainment is -.005. The Sig. for the correlation is .970,
indicating that the correlation occurred solely by chance as it is not
significant. The educational attainment of the husbands and their wives is
306
not meaningfully correlated as the correlation coefficient is very close to
zero, -.005 precisely.
The third table in Output 5.1, Paired Samples Test, displays the Paired
Differences between Husband and Wife to be 8.93939 Mean, 17.89962
Standard Deviation, 2.20329 Standard Error of Mean, 4.53912 Lower
and 13.33967 Upper 95% Confidence Interval of the Difference, 4.057 t,
65 df (degrees of freedom) and Sig. (2-tailed) of .000 that is read as less
than .001 or less than .0005. The t of 4.057 is a function of the Paired
Mean Differences between husband and wife’s educational attainment
(8.93939) divided by the Paired Differences Standard Error Mean
(2.20329). The t of 4.057 is statistically significant because the p-value
(Sig. [2-tailed]) of .000, read as less than .001 or less than .0005, is smaller
than the chosen alpha levels of .05 and of .01. Therefore, the null
hypothesis that “husbands and their wives do not differ significantly in
educational attainment” is rejected in favor of the husbands who have a
higher mean. In other words, the alternate hypothesis that “husbands and
their wives significantly differ in educational attainment” is sustained as the
husbands have a statistically significant higher mean (107.8788) than their
wives’ mean (98.9394) at the 65 degrees of freedom and .05 as well as .01
chosen alphas. Synoptically, spouses’ educational attainment (Paired
Differences Mean of 8.93939; Paired Differences Standard Deviation of
17.89962; and Paired Differences Std. Error Mean of 2.20329) is
statistically significant as t(65) = 4.057, p < .01, d = .499.
It must be recalled that whenever the Lower and Upper 95%
Confidence Interval of the Difference both fall either above zero (0) or
below zero (0), the pair of means under comparison statistically differ
significantly. That is, the mean difference between the two means is
significant. In the analysis that the output is being interpreted here, both the
Lower and Upper 95% Confidence Interval of the Difference (4.53912 and
13.33967, respectively) fall above zero. Therefore, the Mean Difference of
8.93939 is statistically significant. On the contrary, whenever the Lower
95% Confidence Interval of the Difference falls below zero and the Upper
95% Confidence Interval of the Difference falls above zero, the pair of
means under comparison do not significantly vary (differ). The mean
difference in the scenario that zero lies between the Lower and Upper 95%
Confidence Interval of the Difference is not significant and is totally a mere
function of chance. Haven found that there is a significant difference in the
307
educational attainment of husband and wife, what is next to do is
determination of whether the difference has a trivial, small, medium, or
large Effect Size.
Where M1 and M2 are the respective means for the two paired samples
(husband and wife in this case). The numerator of the equation, M1 – M2, is
termed Paired Differences Mean and could as well be abbreviated as
PDM in Paired-Samples t Test. The PDS is the Paired Differences
Standard Deviation. The Paired Samples Test table in the Output 5.1 has
revealed that while the Paired Differences Mean is 8.93939, the Paired
Differences Standard Deviation is 17.89962. Therefore, the Effect Size is
as given in Equation 5:1.
308
Since the calculated t value in the current example is already known, the
d as the measure of Effect Size can as well be computed directly from the t
value of the Paired-Samples t Test. From this perspective, d (the effect size)
can be got as the calculated t value multiplied by the square root of one
divided by the number of paired cases. That is as given in Equation 5:2.
Step 6. View the output with the aid of the left, right, up and down-pointed
arrows at the bottom edge and right edge of the screen.
Step 7. Save the Output 5.1 as Chapter 5 Output 1.spv (see Figure 5.8).
Click the Save this document tool in the Toolbar to have its dialog box
in which you type Chapter 5 Output 1.spv in the File name field and
press Enter or click Save.
309
Figure 5.8 Save Output as Chapter 5 Output 1.spv
Step 8. Print the output. Click the Print tool in the Toolbar to have its
dialog box in which you click OK to get one copy of All visible output
printed as shown in Figure 5.9. The operation can easily be repeated to have
the required number of copies. Alternatively, increase the Number of
copies in the dialog box to the needed number before clicking the
OK so that all the required copies get printed at once.
310
Figure 5.9 Print Output dialog box
Step 9. Exit IBM SPSS by closing all opened SPSS windows. You can also
exit the SPSS by clicking File and in the File dropdown, clicking Exit as
indicated in Figure 5.10.
311
Figure 5.10 File Exit
Step 10. Shut down the system via Start Power Shut down.
312
Paired-Samples t Test on the Spouses’ Educational Attainment can very
swiftly be executed by merely typing Syntax 5.1 into the IBM SPSS
Statistics Syntax Editor and clicking Run All.
To perform the analysis, you must follow the basic SPSS Statistics data
analysis protocols of:
Starting the computer
313
Figure 5.11 Open Data menu
On clicking Open, the Data View Editor with the saved dataset appears
exactly as illustrated in Figure 5. 3.
Click Window in the Menu Bar and in the Window dropdown menu,
click Go to Designated Syntax Window as illustrated in Figure 5.12 to
activate the SPSS Syntax Editor that is shown in Figure 5.13. Another way
of opening the IBM SPSS Statistics Syntax Editor is by clicking File,
selecting New, and clicking Syntax.
314
Figure 5.13 IBM SPSS Statistics Editor
Most carefully enter the Syntax given in Syntax 5.1 as shown in Figure
5.14.
Figure 5.14 IBM SPSS Statistics Syntax Editor with the Paired-Samples t Test Syntax
Next, click Run All. That is, click Run in the Menu Bar of the
IBM SPSS Statistics Syntax Editor, and in the Run dropdown menu,
click All. With this, the entire analysis gets completed and the results
produced in the IBM SPSS Statistics Viewer as displayed in Output 5.2.
315
This Output 5.2 is exactly as Output 5.1. Therefore, the same
interpretation provided for Output 5.1 suffices here. Go through the
interpretation of Output 5.1 again and apply it to the Output 5.2.
Scroll with the aid of the relevant arrows and view the Output 5.2. Save
the Output 5.2 by clicking the Save this document tool in the Toolbar
to have its menu in which you type the document name (Chapter 5 Output
2.spv) into the File name field and press the Enter key on the Keyboard or
click Save as demonstrated in Figure 5.15.
316
Figure 5.15 Save Output As Chapter 5 Output 2.spv
Also save the Syntax by activating the IBM SPSS Statistics Syntax
Editor with the Syntax in question, and click the Save this document icon
in the Toolbar for its menu to appear. In the Save Syntax As dialog
box, type in the name that is suitable for the Syntax, Chapter 5 Syntax
1.sps in this example, as exemplified in Figure 5.16), and press Enter.
317
Figure 5.16 Save Syntax As… Chapter 5 Syntax 1.sps
Print the Output 5.2 by merely clicking the Print tool in the
Toolbar and pressing Enter on the Keyboard or click OK in the dialog box
to have one copy of the All visible output printed.
Exit SPSS by closing all IBM SPSS Statistics windows that are open on
your screen.
Shut down your computer. To do this, click Start Power
Shut down.
SELF-ACTUALIZATION IN 3-YEAR
UNDERGRADUATE PROGRAMMES
One more example on execution of Paired-Samples t Test will definitely
guarantee mastery of how IBM SPSS Statistics is used to perform Paired-
Samples t Test. The data on Self-Actualization were collected from
undergraduate students in four Universities in the United Kingdom when
318
they were admitted in 2014 and when they graduated in 2017. It was a
longitudinal study that aimed at ascertaining Self-Actualization of young
men and women at the point of admission into, termed Admission Self-
Actualization (AS_A), and their Self-Actualization on graduation from,
termed Graduation Self-Actualization (GS_A), 3-year undergraduate
academic programs in four universities in the United Kingdom.
A 5-point 40-item Self-Actualization Scale (S-AS) was administered to
and retrieved from the participants via email in 2014 and 2017, respectively.
The S-AS elicited each participant’s motive to understand oneself, realize
one’s latent potentials, and establish oneself as a total person. While the
minimum possible score that a respondent could get was 40, the maximum
possible score that a respondent could get was 200 as each item on the
Scale was scored 1 to 5 in a positive direction. To ensure equality of length
of time spent in university for the first degree and to prevent unnecessary
differential maturation, students admitted into fast-track 2-year degrees and
those into 4-year sandwich programs were not eligible for the study. Each
participant had a pair of scores (Self-Actualization on Admission and Self-
Actualization on Graduation). The scores are presented in Table 5.2.
319
11 132.00 114.00
12 152.00 120.00
13 172.00 134.00
14 122.00 96.00
15 112.00 100.00
16 172.00 130.00
17 102.00 80.00
18 132.00 102.00
19 172.00 128.00
20 172.00 126.00
21 198.00 131.00
22 192.00 140.00
23 92.00 194.00
24 102.00 100.00
25 127.00 106.00
26 140.00 108.00
27 162.00 130.00
28 172.00 120.00
29 122.00 100.00
30 142.00 100.00
31 132.00 88.00
32 122.00 90.00
33 152.00 109.00
34 152.00 120.00
35 102.00 100.00
36 152.00 130.00
320
37 132.00 100.00
38 112.00 98.00
39 122.00 108.00
40 124.00 90.00
41 138.00 105.00
42 156.00 122.00
43 172.00 140.00
44 182.00 144.00
45 112.00 86.00
46 117.00 100.00
47 112.00 90.00
48 112.00 80.00
49 108.00 78.00
50 110.00 104.00
51 112.00 110.00
52 130.00 112.00
53 132.00 116.00
54 122.00 120.00
55 112.00 102.00
56 162.00 130.00
57 142.00 122.00
58 146.00 125.00
59 130.00 102.00
60 122.00 100.00
61 132.00 102.00
62 122.00 120.00
321
63 122.00 130.00
64 102.00 100.00
65 92.00 88.00
66 122.00 108.00
67 122.00 100.00
68 152.00 120.00
69 132.00 110.00
Research question
What is the Self-Actualization on Admission (AS_A) into and Self-
Actualization on Graduation (GS_A) from 3-year undergraduate programs
in the UK as measured by the mean and Standard Deviation?
Alternate hypothesis
There is a significant difference between Self-Actualization on Admission
(AS_A) into and Self-Actualization on Graduation (GS_A) from 3-year
undergraduate programs in the UK.
Null hypothesis
There is no significant difference between Self-Actualization on Admission
(AS_A) into and Self-Actualization on Graduation (GS_A) from 3-year
undergraduate programs in the UK.
322
Start IBM SPSS Statistics
Figure 5.17 Variable View with the names and labels entered
Get back to the Data View by clicking Data View at the bottom extreme
of the Variable View. This click takes you to the Data View with the
names (GS_A and AS_A) of the variables as exhibited in Figure 5.18.
323
Figure 5.18 Chapter 5 Table 2 Data.sav
324
Figure 5.18 Continued
325
Figure 5.18 Continued
Save the data by clicking the Save this document tool in the Toolbar. In
its dialog box, type a suitable name for the document (Chapter 5 Table 2
Data.sav) in this example as shown in Figure 5.19.
326
Figure 5.19 Save Data As… Chapter 5 Table 2 Data.sav
327
On clicking the Paired-Samples T Test, its dialog box appears as shown
in Figure 5.21. It is in this dialog box that the statistical operations for the
analysis is done.
328
Figure 5.22 Paired-Samples T Test dialog box with variables moved right
At this point, all the necessary statistical operations for the analysis have
been done. So, move cursor to and click OK for IBM SPSS Statistics to
instantly produce the results in the IBM SPSS Statistics Viewer as
presented in Output 5.3
329
INTERPRETATION OF OUTPUT 5.3: SELF-
ACTUALIZATION
The Paired Samples Statistics table of the Output 5.3 shows that
Graduation Self-Actualization (GS_A) has a Mean of 135.0870, Standard
Deviation of 25.13595 and 3.02601 Standard Error of Mean with 69 as the
N (number of cases). Admission Self-Actualization (AS_A) has a Mean of
110.6812 that is lower than that of GS_A. The AS_A has 18.91194
Standard Deviation and 2.27673 Standard Error of Mean. The descriptive
statistics provides answer to the research question of “What is the Self-
Actualization on Admission (AS_A) into and the Self-Actualization on
Graduation (GS_A) from 3-year undergraduate programs in the UK as
measured by the mean and standard deviation?”
The Paired Samples Correlations table indicates a correlation
coefficient of .562 that is statistically significant between the participants’
self-actualization at the point of admission and at graduation time. Those
with high, medium and low self-actualization on admission respectively
maintained relatively high, medium and low self-actualization on
graduation. However, this relationship is not the main concern of the current
330
study. It is the third table of the Output 5.3, Paired Samples Test that is the
core issue of this investigation.
The Paired Samples Test table of the Output 5.3 provides the results for
the hypothesis testing. The Mean Difference between Graduation Self-
Actualization (GS_A) and Admission Self-Actualization (AS_A) is
24.40580. The Paired Differences between GS_A and AS_A has a
Standard Deviation of 21.33577, a Standard Error of Mean of 2.56853,
with 19.28038 Lower and 29.53121 Upper 95% Confidence Interval of
the Difference. The computed t is 9.502 that is statistically significant even
at .001 alpha as the p-value (Sig [2-tailed]) .000, normally written as less
than .0005, is smaller than .001 at 68 degrees of freedom (df). Since the p-
value is less than the chosen alpha, the null hypothesis that “there is no
significant difference between Self-Actualization on Admission (AS_A)
into and Self-Actualization on Graduation (GS_A) from 3-year
undergraduate programs in the UK” is rejected. The rejection is in favour of
Self-Actualization on graduation that has significantly higher mean
(135.0870). In other words, the alternate hypothesis that “there is a
significant difference between Self-Actualization on Admission (AS_A)
into and Self-Actualization on Graduation (GS_A) from 3-year
undergraduate programs in the UK” is sustained in favour of graduation
self-actualization. Kpolovie, Joe and Okoto (2014) had found similar results
in the developing world.
It must be reiterated emphatically that whenever the Lower and Upper
95% Confidence Interval of the Difference both fall either above zero or
below zero, the pair of means under comparison statistically differ
significantly. That is, the Mean Difference between the two means is
significant. In the analysis that the output is being interpreted here, both the
Lower and Upper 95% Confidence Interval of the Difference (19.28038 and
29.53121, respectively) fall above zero. Therefore, the Paired Differences
Mean of 24.40580 is statistically significant.
EFFECT SIZE, d
After ascertaining that there is a significant difference between self-
actualization at graduation from and at admission into the 3-yeear first
degree programme, it becomes very necessary for establishment of whether
331
the difference is a “trivial”, “small”, “medium”, or “large” one. To do this,
Effect Size must be computed. Cohen’s d is the measure of Effect Size in
Paired-Samples t Test. When the results are subjected to d, an Effect Size
of 1.144 is got. The Effect Size of 1.144 is interpreted as a “Large Effect
Size” in accordance with the recommendations by Cohen (1988) and
Cumming (2012).
Effect Size (ES), measured with Cohen’s d in Paired-Samples t Test is
the Paired Differences Mean (24.40580) divided by the Paired
Differences Standard Deviation (21.33577). When the division is done,
the resultant d is 1.144. A d (Effect Size) as high as 1.144 falls within
“Large Effect Size” in the classical recommendations by Cohen (1988) that:
Less than .200 is Trivial ES
Between .200 and .499 is Small ES
Between .500 and .799 is Medium ES
From .800 and above is Large ES.
Perhaps, stating the computation of d as the measure of Effect Size in
Paired-Samples t Test in equation form could allow for easier
comprehension by certain persons. Thus, the Effect Size, d, can be given as
in Equation 5:3.
Where Mpd is the Mean of the Paired Differences, Spd is the Standard
Deviation of the Paired Differences. Note that in the current example, the
Mean of the Paired Differences is 24.40580 and the Standard Deviation of
the Paired Differences is 21.33577. Substituting the symbols in the
Equation 5:3 with their actual values, the One-Sample t Test Effect Size,
measured with Cohen’s d becomes:
332
(the sample size). In equation form, computing d directly from the t value
can thus be given as in Equation 5:2.
333
Save the output as Chapter 5 Output 3.spv
Print output
334
Start IBM SPSS Statistics
Retrieve the data since the entry has already been done and saved as
Chapter 5 Table 2 Data.sav
Window
Run
All.
That is, start the computer, launch IBM SPSS Statistics, and retrieve the
file that has been saved as Chapter 5 Table 2 Data.sav. The retrieval can
simply be done by clicking Open data document icon in the Toolbar
of SPSS Data Editor to get its menu in which you look for and click the
file name (Chapter 5 Table 2 Data.sav) and press Enter as can be seen in
Figure 5.23. Pressing the Enter key on the Keyboard opens the Chapter 5
Table 2 Data.sav as earlier shown in Figure 5.18.
335
Figure 5.23 Open Data menu box
336
as well be got by clicking File, selecting New, and clicking Syntax.
Figure 5.26 IBM SPSS Statistics Syntax Editor with the Syntax
Finally, click Run and All (see Figure 5.27) to have the results displayed
instantly in the SPSS Viewer as presented in Output 5.4. Apart from
clicking Run and All, the results as presented in Output 5.4 could be got by
337
Figure 5.27 Run All
Save the output as Chapter 5 Output 4.spv, and the Syntax as Chapter
5 Syntax 2.sps
338
Print the output
339
Chapter 6
ONE-WAY ANALYSIS OF VARIANCE
Overview
An investigation of a problem that involves simultaneous comparison of
more than two groups demands subjection of the observed data to One-way
Analysis of Variance (ANOVA). ANOVA is a much more powerful
statistical technique than the t Test, and it is very frequently required in
investigations for comparison of three or more groups. Groups to compare
in research and most real-life situations are often more than two for
determination of whether the populations from which they were accurately,
precisely and representatively drawn differ significantly on the dependent
variable. With the down-to-earth treatment of the topic, the user
automatically acquires expert knowledge of SPSS Statistics in ANOVA
execution and of the best possible interpretation of the output. Efficiency of
three types of lightbulb is used for illustration.
340
The ANOVA is a most special statistical technique that best takes care of
the variance within the conditions under investigation while simultaneously
excluding all other sources of variation, such as extraneous, nuisance and
intruding variables from the experiment. ANOVA does this by effectively
comparing the within groups variance (i.e., the variance caused by
extraneous factors like individual differences) with the between groups
variance (i.e., the variance caused by manipulation of the independent
variable). In other words, ANOVA accurately breaks the total variance in an
experimental study, for instance, into two sources that are termed between
groups mean square and within groups mean square (Gray and Kinnear,
2012; Field, 2018; 2013), and divides the former by the latter to arrive at
the F (symbol for ANOVA), which when statistically significant, is the
exclusive effect of the independent variable on the dependent variable.
ANOVA when written without a prefix qualifier like “Two-Way” or
“Within-Subjects”, “Paired-Samples” or “Correlated-Samples”, refers to the
One-Way ANOVA. The One-Way ANOVA refers to One-Way Between-
Subjects ANOVA. It is clear that the ANOVA currently under consideration
is the One-Way Between-Subjects Analysis of Variance. In addition to the
assumptions for comparison of means (normal distribution of the
population, independence of observations and homoscedasticity of
variance) that were discussed in Chapter 3, ANOVA demands that:
i. there is only one independent variable that has different levels,
groups or categories;
ii. there is only one dependent variable on which data are collected at
the interval or ratio scale of measurement;
iii. the cases in each group must be independent of those in other groups
and be drawn from a totally unique population.
Data from an investigation on the effect of types of light bulb on
luminous efficacy are used to illustrate execution of ANOVA with IBM
SPSS Statistics.
341
power consumption, and measured in lumens per watt. It is otherwise
known as lighting efficiency that is the ability of a light bulb to actually
emit visible light using a given amount of power. Three types of bulb were
used:
i. Compact Fluorescent Lamp (CFL): Produces light when an electric
arc passes between cathodes to excite mercury and other gases for
emission of radiant energy that is finally converted to visible light by
a phosphor coating.
ii. Light-Emitting Diode (LED): Is a chemical chip that is embedded
in a plastic capsule for the production of light when voltage is
applied to negatively charged semiconductors that cause electrons to
combine and create a unit of visible light.
iii. Incandescent Bulb (Incan.): Produces light whenever electric
current passes through its filament and causes it to brightly glow.
Two hundred 200 CFL 9W bulbs, 200 LED 8W bulbs and 200
Incandescent 40W bulbs were light up till when the first 50 (i.e., 25%) of
each type of the bulbs went off (could no longer light brightly or only
produces shaking/blinking visibility). The Luminous Efficacy of each of the
150 bulbs during its life time was measured and the data are as shown in
Table 6.1. The three classes or groups of the independent variable (Types of
Bulb) are coded thus: 1 = CFL 9W bulbs, 2 = 8W LED bulbs, and 3 =
40W Incandescent bulbs.
342
44 52 10
46 50 12
32 46 12
36 76 15
44 86 10
50 66 12
32 70 7
42 74 4
35 65 19
43 67 22
34 68 2
51 70 12
47 66 18
49 64 10
29 82 8
31 66 20
30 67 14
51 70 9
40 58 7
44 59 5
43 70 12
34 60 3
44 61 6
61 63 13
53 65 17
54 66 15
343
62 61 10
50 67 18
49 72 20
48 80 21
50 65 8
44 68 10
46 50 14
44 80 19
46 42 17
52 54 19
60 70 14
54 60 12
47 70 4
44 80 12
44 71 9
50 59 9
50 45 10
Research question
What is the luminous efficacy of the three types of bulb (9W CFL, 8W LED
and 40W Incandescent)?
Alternate hypothesis
There is a significant effect of the three types of bulb (9W CFL, 8W LED
and 40W Incandescent) on luminous efficacy.
Null hypothesis
344
Types of bulb (9W CFL, 8W LED and 40W Incandescent) do not have
significant effect on luminous efficacy.
Step 3. Enter the data in the manner that data were entered for performing
Independent-Samples t Test. That is, the entire data on the dependent
variable should be inputted under one column, Var00001, and the three
types of bulb (9W CFL, 8W LED and 40W Incandescent) should be
coded 1, 2 and 3, respectively, under the second column, Var00002. Data in
the example tends to be large. Be fully challenged motivationally to
carefully enter them correctly as every data analysis in real-life situation or
in every investigation that you will be engaged demands entering the data
accurately, which will be large in almost every scenario.
Move the cursor to and click the topmost and extreme left cell in the
Data View of the IBM SPSS Data Editor. Type the data on 9W CFL from
the first value (50) and press Enter, the second value (64) and press Enter,
the third (38) and press Enter, until the last value (50) and press Enter.
Then move cursor to the topmost cell next to the one that you have entered
data, click it and type 1 and press Enter, 1 and press Enter, until where the
data on 9W CFL ends. Press the left arrow on the Keyboard once and click
to ensure that the cursor is in the cell just after the one containing the last
value that you entered on 9W CFL; and continue with entering data on the
8W LED. Type 56 and press Enter, 66 and press Enter, until you type and
press Enter for the last value (45) on 8W LED. Move the cursor to the cell
in Var00002 just below the one that you entered the last 1 code, click it;
and start typing 2 and pressing Enter until the point that the last value on
8W LED is entered. Put cursor on and click the cell just below the one that
345
the last value on 8W LED is entered and type the values for 40W
Incandescent. Type 11 and press Enter, 13 and press Enter, 9 and press
Enter, until you enter the last value (10) and press Enter. Take cursor to
the cell that is just below where the last code of 2 is, click; and type 3 and
press Enter, 3 and press Enter, till you reach the end of all values on 40W
Incandescent.
To provide the actual name, label and value labels for the variables,
move cursor to and click Variable View at the bottom left corner of the
Data View to have the Variable View as shown in Figure 6.1.
Click the first cell under Name that is currently bearing VAR00001 and
type the dependent variable name, LumEfficacy, and in the first cell under
Label, enter Luminous Efficacy, and in the first cell under Measure, click
Unknown and select Scale. Click the second cell under Name, where is
currently called VAR00002, and type BulbTypes; and in the second cell
under Label, enter Types of Bulb, and click Unknown and select Nominal
in the second cell under Measure. Click None in the second cell under
Values and click the small blue dotted box like ellipse in it for a Value
Labels dialog box to appear as shown in Figure 6.2.
346
Figure 6.2 Value Labels dialog box
In the box beside Value, enter 1 and in the box beside Label, enter 9W
CFL and click Add to have Figure 6.3.
Figure 6.3 Value Label dialog box with 1 and 9W CFL added
Type 2 in the box beside Value, and enter 8W LED in the box beside
Label, and click Add to have Figure 6.4.
347
Figure 6.4 Value Labels with 2 and 8W LED added
In the box beside Value, type 3 and enter 40W Incandescent in the box
beside Label and click Add as indicated in Figure 6.5.
Click OK to get back to Figure 6.1. In it, click Data View to see the
Data View with all the entered data and the variable names as presented in
Figure 6.6.
348
Figure 6.6 Chapter 6 Table 1 Data.sav
349
Figure 6.6 Continued
350
Figure 6.6 Continued
351
Figure 6.6 Continued
352
Figure 6.6 Continued
353
Figure 6.6 Continued
354
Figure 6.6 Continued
Step 4. Save the data. Move cursor to the Toolbar and click the Save this
document icon for the Save Data As… dialog box to appear as shown
in Figure 6.7. In it, type the name of the document, Chapter 6 Table 1
Data.sav, in the box beside File name and click Save or press Enter.
355
Figure 6.7 Save Data As… Chapter 6 Table 1 Data.sav
356
A click on the One-Way ANOVA produces the One-Way ANOVA
dialog box in which the entire statistical operations required for the analysis
are done with greatest ease as given in Figure 6.9.
357
Figure 6.10 One-Way ANOVA dialog box with variables transferred
are selected by checking the small boxes beside them for illustration
here as can be seen in Figure 6.11.
358
Figure 6.11 One-Way ANOVA: Post Hoc Multiple Comparisons
Click Continue and it will take you back to the One-Way ANOVA
main dialog box for you to proceed with other analytical operations.
Then, click Options pushbutton to have its dialog box. In the Options
359
360
361
INTERPRETATION OF OUTPUT 6.1: BULB TYPES
EFFECT ON LUMINOUS EFFICACY
There are six parts of the Output 6.1 (five tables and one graph). Each of the
parts is listed and interpreted accordingly. They are Descriptives, Test of
Homogeneity of Variances, ANOVA Luminous Efficacy, Post Hoc Multiple
Comparisons, Homogeneous Subsets Luminous Efficacy, and Mean Plots.
In addition, Effect Size is calculated and interpreted as well for each
significant mean difference.
Descriptives
This section of the Output has most simply outlined the N (number of
cases), Mean, Std. Deviation, Std. Error, Lower Bound and Upper Bound of
95% Confidence Interval for Mean, the Minimum score and Maximum
score, respectively, for each of the three types of light bulb (9W CFL, 8W
362
LED and 40W Incandescent). The same statistical descriptions are also
given for the Total 150 bulbs that were observed and measured. For
instance, each of the three types of bulb has 50 cases; the Mean and Std.
Deviation respectively for 9W CFL bulbs are 44.9800 and 8.65316, for 8W
LED bulbs are 66.0000 and 10.12372, for 40W Incandescent bulbs are
12.0000 and 5.03863, and for the Total are 40.9933 and 23.74741. The
other columns of this table can easily be interpreted in like manner against
the rows. The means and standard deviations in the Descriptives table are
answers to the posed research question.
363
.05 and even of .01, the null hypothesis that “the three types of bulb (9W
CFL, 8W LED and 40W Incandescent) do not have significant effect on
luminous efficacy” is rejected. Therefore, the alternate hypothesis that
“there is a significant effect of the three types of bulb (9W CFL, 8W LED
and 40W Incandescent) on luminous efficacy” is upheld. Rejection of the
omnibus null hypothesis demands that Post Hoc Multiple Comparisons be
done to ascertain the particular pairs of means that differ significantly,
which is reported in the next two parts of the interpretation of Output 6.1.
364
Homogeneous Subsets Luminous Efficacy
This table has shown that each of the three types of bulb has a luminous
efficacy mean that is significantly different from the others. For instance,
the Ryan-Einot-Gabriel-Welsch Range has revealed beyond all possible
doubts statistically that the 40W Incandescent bulbs have luminous efficacy
mean of 12.0000 that is overwhelmingly lower than the luminous efficacy
mean of 9W CFL bulbs (44.9800), which is in turn absolutely lower than
the luminous efficacy mean of the 8W LED bulbs (66.0000). Each of the
mean differences is statistically significant as no two means fall under the
same homogenous subset.
Means Plots
The Means Plots has graphically expressed the relative positions of the
Types of Bulb (9W CFL, 8W LED and 40W Incandescent) along the
horizontal axis against the Mean of Luminous Efficacy at the vertical axis.
This is a pictorial description of the significant effect of the Types of Bulb
on Luminous Efficacy to show that the 40W Incandescent is incomparably
less effective than 8W LED, and the 9W CFL. Furthermore, the 8W LED
has outstandingly more luminous efficacy than the 9W CFL bulbs. The
interpretation cannot be complete without presentation of the Effect Size
that is addressed in the final section.
365
In the ANOVA Luminous Effect table of the Output 6.1, the Sum of
Squares Between Groups is 74092.013 and the Sum of Squares Total is
84026.993. Therefore, the Eta-Squared is as follows.
The conventions for trivial, small, medium, and large Effect Sizes as
measured with Eta-Squared (ɳ2) for ANOVA are .000 to .009, .010 to
.059, .060 to .139, and .140 and above, respectively, in accordance with the
recommendations of Cohen (1988) and Cumming (2012). That is,
determination of Effect Size in ANOVA is measured with Eta Squired,
and the Eta Squired (ɳ2) is interpreted as follows:
.000 - .009 = Trivial Effect Size
.010 - .059 = Small Effect Size
.060 - .139 = Medium Effect Size
.140 and above = Large Effect Size.
Therefore, the computed Eta-Squared of .8817, .882 approximately as
the effect size is better stated in three decimal places, for the null hypothesis
that has just been tested with regard to effect of Types of Bulb on
Luminous Efficacy, is a marvellously Large Effect Size. This confirms the
rejection of the null hypothesis beyond every statistical, reasonable and
practical doubts to absolutely indicate that Types of Bulb does not only
have significant effect on Luminous Efficacy; but that the effect it has on
Luminous Efficacy is also an overwhelmingly large effect size.
Furthermore, Eta-Squared can be expressed in terms of the percentage
of variance in the dependent variable that is accounted for by the
independent variable. This is done by multiplying the Eta-Squared by 100
as given in Equation 6:2.
366
Where SSB is the Sum of Squares Between Groups and SST is the Sum
of Squares Total. Therefore, the Effect Size of 88.176, written as .882 is
wonderfully large in practice as it indicates that Types of Bulb (the
independent variable) accounts for as high as 88.2% of the variance in
Luminous Efficacy (the dependent variable).
367
Standard Deviation. Therefore, d is easily computed as provided in
Equation 6:3.
368
The d of 2.232 is a very large Effect Size in line with the
recommendations for interpretation of d as a measure of effect size that I
have much earlier given in Table 3.2 of Chapter 3. But for time saving and
instant reference, let me restate it here.
Recommendations for Interpretation of d in ANOVA Pairwise
Comparisons:
d AS EFFECT SIZE WITH DESCRIPTIVE CLASSIFICATION
0.000 to 0.199 = Trivial Effect Size
0.200 to 0.499 = Small Effect Size
0.500 to 0.799 = Medium Effect Size
0.800 and above = Large Effect Size
Where the groups are not equal in the sample size, d can still be used as
the measure of effect size, but the equation for it will rather be as given in
Equation 6:4.
369
Where d is the effect size of the pairwise means comparison; M1 is the
mean of the first group; M2 is the mean of the second group; N1 is the
number of cases in the first group; N2 is the number of cases in the second
group; S1 is the standard deviation of the first group; S2 is the standard
deviation of the second group; S12 is the variance of the first group; and S22
is the variance of the second group.
For the pairwise comparison of the means of 8W LED (M1 = 66.0000,
S1 = 10.12372) and 9W CFL (M2 = 44.9800, S2 = 8.65316), that has N1 =
50 and N2 = 50; the d is as in Equation 6:4.
370
Effect Size, d, for 9W LED versus 40W Incandescent
Bulbs
Since the two groups are equal in sample size, (N = 50 in each group), the d
as a measure of Effect Size in ANOVA Post Hoc Tests Multiple Pairwise
Comparisons can be computed easily thus as given in Equation 6:3.
371
The Luminous Efficacy of the 9W CFL bulbs has a Mean (M1) of
44.9800 and a Standard Deviation (S1) of 8.65316. The Luminous Efficacy
of the 40W Incandescent bulbs has a Mean (M2) of 12.0000 and a Standard
Deviation (S2) of 5.03863. Therefore, the Effect Size, d, is:
372
The Pairwise Comparison of Means of the 9W CFL (M1 = 44.9800,
SD1 = 8.65316) and 40W Incandescent (M2 = 12.000, SD2 = 5.03863)
with this equation that is suitable for unequal sample size, the Effect Size, d,
is as follows.
Yes, the computed Effect Size, d of 4.658 is exactly the same with when
Equation 6:3 was used because the two groups in the example under
comparison have equal sample size of 50. So, whenever the sample sizes
are equal or about equal, just make use of the Equation 6:3. When the two
comparison groups differ in sample size, use Equation 6:4.
373
Effect Size, d, for 8W LED versus 40W Incandescent
Bulbs
The Pairwise Comparison of the Means of the 8W LED (M1 = 66.000, SD1
= 10.12372) and 40W Incandescent (M2 = 12.000, SD2 = 5.03863) when
the pairwise groups have equal sample size (N = 50 for each group) can
have the d, Effect Size, calculated as in Equation 6:3.
374
Supposing that the sample size in the two groups are different, the Effect
Size could still be computed with d, but with the Equation 6:4.
Since the 8W LED bulb has M1 of 66.0000 and S1 of 10.12372 while the
40W Incandescent bulb has M2 of 12.0000 and S2 of 5.03863, the d in
accordance with the current equation will be:
375
In a most apt manner, the entire Output 6.1 can technically be written thus.
The effect of Types of Bulb on Luminous Efficacy, F(2, 147) = 548.140, p
< .001, ɳ2 = .882 (a marvellously large effect size). Scheffe, Bonferroni
and Ryan-Einot-Gabriel-Welsch Range Post Hoc Multiple Comparisons
tests indicate that 8W LED (M = 66.000, SD = 10.12372) has significantly
preponderant Luminous Efficacy than 9W CFL (M = 44.9800, SD =
8.65316), d = 2.232 (a very large effect size). In turn, the 9W CFL (M
=44.9800, SD = 8.65316) has significantly preponderant Luminous Efficacy
than 40W Incandescent (M = 12.0000, SD = 5.03863), d = 4.658 (an
exceptionally large effect size). The 8W LED (M= 66.0000, SD =
10.12372) has more overwhelmingly significant Luminous Efficacy than
the 40W Incandescent (M = 12.0000, SD = 5.03863), d = 6.753 (a
profoundly large effect size).
Either of one 8W LED bulb or one 9W CFL bulb produces
overwhelmingly greater brightness than one 40W Incandescent bulb. Yet,
the amount of wattage required to light one 40W Incandescent bulb can be
used to light as many as five 8W LED bulbs. Similarly, the wattage
demanded to light one 40W Incandescent bulb can be used to light more
than four 9W CFL bulbs. Furthermore, while 40W Incandescent bulb has
1000 hours lifespan on the average, 9W CFL bulb has an average of 10,000
hours lifespan, and the lifespan of 8W LED bulb is 25,000 hours on the
average. Therefore, the 8W LED bulb is indeed incomparably better than
9W CFL bulb, which is in turn, incomparably better than the 40W
Incandescent bulb. In addition, while the 40W Incandescent bulb generates
excessive heat as a by-product, 8W LED bulb and 9W CFL bulb produce
minimal heat.
It is little wonder that the Penn State Center for Nanoscale Sciences
(2016) after displaying and observing three types of bulbs in one
investigation, stated that “three bulbs in the display each have a light output
of 400 lumens, but they require different amounts of power. The
incandescent bulb uses 60 W, the fluorescent bulb uses 7 W, and the LED
bulb uses 6.5 W.”
Step 6. View the output by scrolling with the aid of the right, left, up, and
down-pointed arrows at the extreme bottom and extreme right edge of your
screen.
376
Step 7. Save the output by clicking Save this document tool on the
Toolbar and type the document name, Chapter 6 Output 1.spv in the box
beside File name and click Save or press Enter as given in Figure 6.13.
Step 8. Print the output by clicking the Print icon on the Toolbar to
have the Print dialog box; when no part of the output is selected, click OK
for All visible output as shown Figure 6.14.
377
Figure 6.14 Print All visible output
Follow similar procedure to print the data too. Note that printing of the
data has to be done when the Data View of the Data Editor containing the
requisite data is open.
Step 9. Exit IBM SPSS Statistics by closing all opened SPSS windows.
378
To perform the ANOVA with the use of IBM SPSS Statistics Commands
Syntax:
Switch on the computer
Retrieve the data since the entry has already been done and saved as
Chapter 6 Table 1 Data.sav
Window
379
Run
All
That is, start the computer, open IBM SPSS Statistics, and retrieve the
dataset file that has been saved as Chapter 6 Table 1 Data.sav. The
retrieval can simply be done by clicking Open data document icon in
the Toolbar of SPSS Data Editor to get its menu in which you look for, and
click the file name (Chapter 6 Table 1 Data.sav), and press Enter as
illustrated in Figure 6.15. Pressing the Enter key on the Keyboard opens
the Chapter 6 Table 1 Data.sav as earlier shown in Figure 6.6.
In the Open Data menu box, activate the IBM SPSS Statistics Syntax
Editor by clicking Window and clicking Go to Designated Syntax
Window as illustrated in Figure 6.16. The IBM SPSS Statistics Syntax
380
Editor can as well be opened by clicking File in the Menu Bar, selecting
New and clicking Syntax.
381
Finally, click Run and click All as shown in Figure 6.19 to have the
output displayed instantly in the IBM SPSS Statistics Viewer as presented
in Output 6.2.
382
383
384
INTERPRETATION OF OUTPUT 6.2: ONE-WAY
ANOVA LUMINOUS EFFICACY
The Output 6.2, arrived at via application of IBM SPSS Statistics Syntax, is
exactly like Output 6.1 that was reached through SPSS Dialog Boxes Point
and Click method. Therefore, exactly the same explanation that was
proffered for Output 6.1 is applicable with Output 6.2. Endeavour to go
through the interpretation offered for Output 6.1 again and replicate it with
the Output 6.2. It will enable you to better grasp and master explanation of
how One-Way ANOVA results are presented and elucidated.
Proceed with the remaining routine protocols of how IBM SPSS
Statistics is used for data analysis to finally shut down your system.
View the output
385
Save the output as Chapter 6 Output 2.spv, and the Syntax as Chapter
6 Syntax 1.sps
386
Analyze, using the Univariate Method.
Now that you have activated the SPSS Data Editor containing the data
file called Chapter 6 Table 1 Data.sav, the analysis proper is what has to
be done. However, if you are yet to retrieve the file and is not very sure of
how to do it, move cursor to the Menu Bar and select File Open
Data (see Figure 6.20), and the Open Data dialog box will appear.
From the Open Data dialog box, as can be seen in Figure 6.21, click the
file name, Chapter 6 Table 1 Data.sav. This click highlights the name of
the file. Then, click Open in the dialog box or press Enter on the Keyboard
to have the data file as shown earlier in Figure 6.6 on your screen.
There is also another way to quickly retrieve the saved data file. While
in the IBM SPSS Statistics Data Editor, merely clicking Open data
document icon in the Toolbar automatically produces the Open Data
dialog box in Figure 6.21 where you select the file name, Chapter 6 Table
1 Data.sav, and click Open or press Enter for the data document to open.
387
Figure 6.21 Open Data dialog box
Now that the dataset has opened on your screen as in Figure 6.6,
perform the analysis by clicking Analyze General Linear Model
Univariate as demonstrated in Figure 6.22.
388
On clicking Univariate, the main Univariate dialog box appears as
illustrated in Figure 6.23. It is in this Univariate main menu that the entire
statistical operations required for execution of the ANOVA are performed.
389
Figure 6.24 Univariate dialog box with the variables moved from the left box
Click Plots pushbutton to have its dialog box. In the Univariate: Profile
Plots dialog box, BulbTypes is in the Factors box at the left and is already
highlighted. So, click the arrow that is directly pointing at Horizontal
Axis. With this, the BulbTypes appears in the Horizontal Axis box at the
right as given in Figure 6.25.
390
Figure 6.25 Univariate: Profile Plots with BulbTypes moved into Horizontal Axis
Move cursor to, and click Add for the plots operation to be completed as
demonstrated in Figure 6.26. You can observe that the BulbTypes has
disappeared from the Horizontal Axis panel and appeared in the field just
below Plots. You can equally see that Line Chart and Bar Chart have
been activated for you to choose one, though Line Chart is the default in
SPSS. Click Bar Chart for it to be used in the graphical representation of
the magnitude of the dependent variable’s (Luminous Efficacy) means as a
function of the independent variable (BulbTypes) in the eventual output.
391
Figure 6.26 Univariate: Profile Plots with Bar Chart checked
Click Continue to return to the main Univariate dialog box. Next, click
Post Hoc pushbutton for the Univariate: Post Hoc Multiple Comparisons
for Observed Means dialog box to be displayed as shown in Figure 6.27.
392
Figure 6.27 Univariate: Post Hoc Multiple Comparisons for Means menu
Click the arrow pointing at the Post Hoc Tests for box at the right
for the highlighted BulbTypes in the Factor(s) to be transferred into the
Post Hoc Tests for panel. In this dialog box, check the checkbox beside
Bonferroni, Scheffe, and R-E-G-W-Q for these Post Hoc Multiple
Comparisons tests to be done as parts of the output to be arrived at. An
example of the Univariate: Post Hoc Multiple Comparisons for
Observed Means when the needful Post Hoc tests checkboxes have been
checked is shown in Figure 6.28. Next, click Continue to proceed with the
analysis.
393
Figure 6.28 Univariate: Post Hoc dialog box with Bulb Types transferred
394
Figure 6.29 Univariate: Options dialog box
395
Figure 6.30 Univariate: Estimated Marginal Mean dialog box
396
Figure 6.31 Univariate: Estimated Marginal Means with variables moved to Display
Means for
Finally, the entire statistical operations required for the ANOVA, using
SPSS Univariate method, have been completely done. So, click OK for
SPSS to produce the results in the IBM SPSS Statistics Viewer as
presented in Output 6.3.
397
Estimated Marginal Means
398
Post Hoc Tests
399
Homogeneous Subsets
400
Profile Plots
401
INTERPRETATION OF OUTPUT 6.3: UNIVARIATE
ANOVA ON BULB TYPES
The results presented in this Output 6.3 are not much different from those
presented and explained in Outputs 6.1 and 6.2 as the means and standard
deviations as well as the sums of squares and mean squares for between
groups and within groups are the same. The computed F in all the scenarios
are equal and the omnibus null hypothesis is rejected with equal Effect Size
in all the three outputs. Each pairwise comparison is statistically significant
with the same Effect Size in all the three outputs. However, for the fact that
the Output 6.3 tends to be more elaborate with more sub-tables, I have
presented a very brief interpretation for the very relevant aspects of the
output as an additional guide.
402
There are nine sections in the Output, eight sub-tables and one graph.
The sections of the Output (Univariate Analysis of Variance) are: Between-
Subjects Factors, Descriptive Statistics, Test of Between-Subjects Effects,
Estimated Marginal Means for Types of Bulb, Pairwise Comparisons,
Univariate Tests, Post Hoc Tests – Types of Bulb Multiple Comparisons,
Homogeneous Subsets – Luminous Efficacy, and Profile Plots – Estimated
Marginal Means of Luminous Efficacy. One or two words of interpretation
is offered for each of these sections; and finally, Effect Size is provided as
necessary.
Between-subjects Factors
This table has presented the Value Labels and N (Number of cases) for the
three types of bulb (1.00 = 9W CFL, 2.00 = 8W LED, and 3.00 = 40W
Incandescent). The N for each type of bulb is 50.
Descriptive Statistics
The Descriptive Statistics section of Output 6.3 has presented the Luminous
Efficacy Mean, Standard Deviation and Number of cases for each of the
three Types of Bulb, and for the Total. The 9W CFL has a Mean of 44.9800
and a Std. Deviation of 8.65316. The 8W LED has 66.0000 Mean and
10.12372 Std. Deviation. The 40W Incandescent has a Mean of 12.0000
and a Std. Deviation of 5.03863. Each of the three Types of Bulb has N of
50. The Total N is therefore 150 with a Total (grand) Mean and Std.
Deviation of 40.9933 and 23.74741, respectively. The Luminous Efficacy
means and standard deviations of the three types of bulb serve as the answer
to the research question.
403
Between-Subjects Effects are the BulbTypes, the Error, and the Corrected
Total. These three rows are critically necessary for One-Way ANOVA
execution by all standards; and are comparable with what other ANOVA
procedures ordinarily yield.
The Type III Sum of Squares for the six rows represent the following:
a. Corrected Model that is the sum of all the between-subjects effects in
the analysis. But since there is only a single between-subjects effect
in the One-Way ANOVA, the BulbTypes in this case, the Corrected
Model sum of squares is bound to be equal to the effect of the
BulbTypes. You may wish to ask why it is called corrected model. It
is termed Corrected Model because its value totally excludes the
effect of the Intercept that is specific to regression analysis.
b. Intercept which is the regression-specific effect represents the value
of the intercept, Y, in regression model.
c. BulbTypes is the effect of the actual independent variable under
investigation (Types of Bulb) on the dependent variable (Luminous
Efficacy) in the study used for illustration of the ANOVA analysis
(see Kpolovie and Ololube, 2020).
d. Error is the Within Groups effect that estimates the measurement
error and all individual differences as well as the influence of all
other nuisance or extraneous variables on the dependent variable. The
Error Mean Square (Mean Square Within Groups) is the final
denominator of ANOVA in order to completely eliminate the
influence of probable extraneous variables from affecting the
dependent variable so that whatever that becomes of the dependent
variable is confidently the effect of the manipulated independent
variable.
e. Total is the total sum of squares including the regression-specific
Intercept effect. It is mandatorily needed in regression analysis; but
not in ANOVA and as such, it is never considered in the
determination of Effect Size in ANOVA when the Univariate
statistical approach is used.
f. Corrected Total refers to the total for the reduced model that includes
only the effects of the independent variable and the error source of
variance. That is, the Corrected Total does not include the effect of
the Intercept at all, and so guarantees that ANOVA is indeed having
only two sources of variation or variance (the Between Groups [the
404
independent variable] and the Within Groups [the Error]).
Consequently, when computing the value of Partial Eta Squared
(ɳp2) that is used for measuring Effect Size in ANOVA that is due
exclusively to the manipulated independent variable, it is the
Corrected Total that is used to constitute the denominator in a
scenario that Univariate Analysis of Variance was the adopted
approach.
While the Sum of Squares for BulbTypes (i.e., the Between Groups)
divided by its df (degrees of freedom) is equal to the Mean Square Between
Groups, the Sum of Squares Error divided by its df is equal to the Mean
Square Error (Within Groups). The Mean Square Between Groups
(37046.007) divided by the Mean Square Error (67.585) is equal to the F
(548.140). This F is very large and is statistically significant even at .001
alpha because the p-value (Sig.) is .000, written as .001 or read as less than
.0005. Therefore, the omnibus null hypothesis that “the three types of bulb
(9W CFL, 8W LED and 40W Incandescent) do not have significant effect
on luminous efficacy” is rejected.
405
variables and by the interaction of all the independent variables. However,
with increase in the number of independent variables in an ANOVA model,
the proportion of variance accounted for by each particular independent
variable becomes slightly less precise in the estimation of the dependent
variance in the population.
Partial Eta Squared is the statistical measure of Effect Size that indicates
both the main and interaction effects of one or more categorical
independent variables on a dependent variable by dividing the Sum of
Squares for the effects by the Sum of Squares for both the effects and the
associated error, such as individual differences in the dependent variable.
This denominator is referred to as Sum of Squares Corrected Total in the
Univariate ANOVA output. The size of the effect of an independent
variable on a dependent or outcome variable cannot be determined by the
probability level (p-value or Sig.) arrived at in the analysis. The magnitude
of the independent variables effect on a dependent variable can only be
determined directly by a special measure that is collectively referred to as
Effect Size. For comparison of two means as done with t Test, Cohen’s d, is
used as a standardized measurement to measure the Effect Size accurately.
For comparison of three or more groups that is computed with ANOVA, Eta
Squared, Partial Eta Squared, and R Squared could be the measure of the
Effect Size.
The Eta Squared in ANOVA is identical with R Squared in Multiple
Regression; and it measures the proportion of the Total Variance in the
dependent variable that is accounted for by the differences among the
means of the treatment groups (the categorical independent variable).
Technically, Eta Squared is the ratio of the Between Groups Sum of Squares
to the Total Groups Sum of squares in ANOVA. When the sample size is
very large, Eta Squared estimates the population variance with regard to the
variables under study without bias. Use of a truly representative sample that
is suitably large with accuracy and precision from each of the populations is
recommended for each investigation in addition to adherence to the
assumptions that underlie comparison of means as presented earlier in
Chapter 3. For a small sample size, not recommended, the square of the Eta
Squared, known in Multiple Regression as Adjusted R Squared, estimates
the population with slightly higher precision. Epsilon Squared or Omega
Squared could also be used for the purpose.
406
Partial Eta Squared, ɳp2
Partial Eta Squared (ɳp2) is very popularly used as a measure of Effect
Size in Univariate ANOVA to overcome the shortcoming in Eta Squared
that addition of more variables to the ANOVA model tends to reduce the
proportion of variance in the dependent variable that is accounted for by
any one independent variable. Also, Partial Eta Squared better allows for
greater precision in the comparison of the effect of one independent
variable across many different studies with the same or similar dependent
variable.
How is Partial Eta Squared (ɳp2) able to accomplish overcoming the
shortcomings in Eta Squared and gain so much popularity, you may ask.
Partial Eta Squared enables an investigator to compare the effect of a
particular treatment variable (independent variable) in different studies that
contain different covariates (other factors) because Partial Eta Squared
eliminates every variation in the dependent variable that is explained or
accounted for by other variables from the denominator. The denominator in
Partial Eta Squared is the sum of the unexplained variation in the dependent
variable and the variation in the dependent variable that is explained by just
the independent variable under study. That is, the denominator in Partial Eta
Squared is not just the total variation in the dependent variable as in Eta
Squared.
Aptly put, Partial Eta squared is the best measure of Effect Size in
ANOVA models, and it is arrived at by dividing the Sum of Squares
Between Groups by the Sum of Squares Between Groups plus the Sum of
Squares Within Groups (i.e., Sum of Squares Error) that are collectively
known in Univariate ANOVA models as Corrected Sum of Squared Total.
In equation form therefore, the Partial Eta Squared can simply be given as
in Equation 6:5.
Where SSB is the Sum of Squares Between Groups, and SSE is the Sum
of Squares Error which is otherwise referred to as Sum of Squares Within
Groups (Kpolovie, 2018).
The denominator of the ɳp2, the Partial Eta Squared, which is Sum of
Squares Between (SSB) plus Sum of Squares Error (SSE) is collectively
407
known and more popularly referred to as Corrected Sum of Squares Total.
In fact, the IBM SPSS Statistics calls it as the Sum of Squares Corrected
Total (SSCT). Substituting this name, Sum of Squares Corrected Total, in
the denominator of the preceding equation, the Partial Eta Squared becomes
as given in Equation 6:6.
Where SSB is the Sum of Squares Between Groups, and SSCT is the
Sum of Squares Corrected Total.
In One-Way ANOVA that has only one independent variable, the Partial
Eta Squared is exactly equal to the Eta Squared because the Sum of Squares
Total as the denominator in Eta Squared is made up of the Sum of Squares
Between Groups plus the Sum of Squares Within Groups, which is the Sum
of Squares Corrected Total in the Partial Eta Squared. This denotes that
either Eta Squared (ɳ2) or Partial Eta Squared (ɳp2) can be used to rightly
establish the Effect Size in One-Way ANOVA as long as Sum of Squares
Corrected Total in the Univariate ANOVA is used as the denominator,
instead of the Sum of Squares Total which also contains the Intercept as a
Source of variation. But in higher models of ANOVA with more than one
independent variables, the denominator has two or more Between Sums of
Squares plus the Error Sum of Squares to better express the proportion of
variance in the dependent variable that is explained solely by each of the
independent variables, and jointly by the interaction effect of the combined
independent variables in the model.
Good a thing, there is a statistical operation in IBM SPSS Statistics that
automatically computes and provides the Partial Eta Squared as part of
the output, when checked in the process of the analysis or included in the
SPSS Univariate ANOVA Syntax. The function is “Estimates of effect
size” that is within the Univariate: Options dialog box. You can recall that
the Estimates of effect size checkbox was checked in Figure 6.29 in the
course of the analysis. Consequently, Partial Eta Squared as the measure
of Effect Size constitutes the last column in the Test of Between-Subjects
Effects table of the Output 6.3. In that column, Corrected Model has .882
Partial Eta Squared, Intercept has .962 Partial Eta Squared, and BulbTypes
has .882 Partial Eta Squared. Types of Bulb (BulbTypes) which is the
independent variable under investigation has overwhelmingly high effect on
Luminous Efficacy with .882 Effect Size (the Partial Eta Squared) that is
408
interpreted as a very Large Effect Size. Recommendations for the
interpretation of Effect Size in ANOVA that is measured with Partial Eta
Squared are the same as given earlier for Eta Squared. The Partial Eta
Squared (ɳp2) is interpreted as:
.000 - .009 = Trivial Effect Size
.010 - .059 = Small Effect Size
.060 - .139 = Medium Effect Size
.140 and above = Large Effect Size.
Perhaps one of the reasons for the popularity of Partial Eta Squared as
the best measure of Effect Size in ANOVA models is because IBM SPSS
Statistics and some other statistical software compute it automatically and
report it as part of the output. So, you do not have to worry on how to
compute Partial Eta Squared with hand calculator. With just a check on the
“Estimates of effect size” checkbox within the Univariate: Options dialog
box, SPSS automatically performs and reports the Partial Eta Squared in the
ANOVA model that is adopted in your work.
409
EFFECT SIZE COMPUTION AS PARTIAL ETA
SQUARED
Effect Size in Univariate ANOVA is measured as Partial Eta Squared
(ɳp2). But do not forget the explanation that I have given that the Partial Eta
Squared in Univariate Analysis of Variance, using IBM SPSS Statistics,
automatically reports Effect Size to be exactly equal to Eta Squared in
ANOVA model that has only one independent (treatment) variable.
Therefore, the Partial Eta Squared associated with the effect of
BulbTypes on Luminous Efficacy can be calculated both with Partial Eta
Squared and with Eta Squared. The Partial Eta Squared is computed as the
Between Groups Sum of Squares (i.e., the BulbTypes Sum of Squares)
divided by the Corrected Total Sum of Squares just as has been discussed
under Output 6.1 before in which Total Sum of Squares was used in place
of the Sum of Squares Corrected Total. In the current output, while the
Between Groups Sum of Squares for BulbTypes is 74092.013, the Sum of
Squares Corrected Total is 84026.993. The Partial Eta Squared as the
Effect Size in the Univariate ANOVA can be given and calculated with this
equation in line with Equation 6:6. The Sum of Squares Between Groups
(SSB) in the current example is the Sum of Squares for BulbTypes. Thus:
This implies overtly that the Effect Size of .882 is wonderfully large in
practice as it indicates that Types of Bulb (the independent variable)
accounts for as high as 88.2% of the variance in Luminous Efficacy (the
dependent variable). If an effect size that is up to .140 is classified as a large
effect size, then this effect size of .882 is not only a large effect size, but a
profoundly large effect size that is very uncommon to find. Computation of
Eta Squared in ordinary One-Way ANOVA that is executed via Compare
Means, the denominator is the Sum of Squares Total, but if the One-Way
410
ANOVA is executed via Univariate Analysis of Variance, the denominator
is the Sum of Squares Corrected Total which harmonizes the Eta Squared
with the Partial Eta Squared. It has been emphasized enough that in
calculation of Partial Eta Squared, the denominator is the Sum of Squares
Corrected Total, and not the Sum of Squares Total as in Eta Squared that the
analysis employed the ANOVA statistical operations available in Compare
Means.
Pairwise Comparisons
The Pairwise Comparisons that is based on the Univariate ANOVA
Marginal Mean Difference, using Least Significant Difference (LSD),
shows a significant difference at .05 alpha for each of the pairwise
comparisons. The significant Mean Difference between 9W CFL and 8W
LED is -21.020, 9W CFL and 40W Incandescent is 32.980; 8W LED and
9W CFL is 21.020, 8W LED and 40W Incandescent is 54.000; 40W
Incandescent and 9W CFL is -32.980, and between 40W Incandescent and
8W LED is 54.000. A Std. Error of 1.644 and a P-valuer (Sig.) of .000 (read
as less than .0005) are indicated for each Pairwise Mean Difference. The
Lower as well as Upper Bounds at 95% Confidence Interval is equally
provided for each Pairwise Mean Difference in the table.
Univariate Tests
The Univariate Tests is a brief summary table that also provides the Sum of
Squares, df, Mean Square, F and Sig. against the Contrast (BulbTypes or
the Between Groups) and the Error (the Within Groups) to clearly show the
F value of 548.140 that depicts a significant effect of Types of Bulbs on
Luminous Efficacy. This F test is based directly on the linearly
independent pairwise comparisons against the estimated marginal means.
411
Post Hoc Tests on Types of Bulb Multiple Comparisons
This table is based on the observed means and the Error term (Within
Groups) Mean Square of 67.585. Multiple Comparisons were done for the
various pairwise means, using Scheffe, Bonferroni and Ryan-Einot-Gabriel-
Welsch Range. It can very clearly be seen from the Multiple Comparisons
table for both Scheffe and Bonferroni that each of the pairwise comparisons
has statistically significant difference as the Mean Difference (I-J) column
for each of the Mean Differences has asterisk (*) that denotes statistically
significant difference at .05 alpha level. A further look shows that indeed
the mean difference between each pair is overwhelmingly preponderant that
it is even significant at as low as .000 (written as .001 or less than .0005) as
indicated in the p-value (Sig.) column. The 95% Confidence Interval Lower
Bound and Upper Bound column has further shown that for each pair of the
means, the mean difference is significant because for every paired means,
the Lower Bound and the Upper Bound are either completely above zero or
completely below zero. Using Scheffe for instance, the luminous efficacy
Mean Difference between 9W CFL and 8W LED bulbs is -21.02000* and
has 1.64420 Std. Error, .000 Sig., plus -25.0859 Lower Bound and -16.9541
Upper Bound at 95% Confidence Interval. The 9W CFL and 40W
Incandescent bulbs have 32.98000* Mean Difference, 1.64420 Std. Error,
.000 Sig., 28.9141 Lower Bound and 37.0459 Upper Bound at 95%
Confidence Interval. In fact, the Luminous Efficacy Mean Difference
between 8W LED bulb and 40W Incandescent bulb is as high as 54.0000*,
with 1.64420 Std. Error, .000 Sig., and 49.9341 and 58.0659 Lower Bound
and Upper Bound, respectively, at 95% Confidence Interval. Similar
interpretation is applicable to each of the paired types of bulb that constitute
a row in this table, irrespective of whether it is Scheffe or Bonferroni’s tests
of Multiple Comparisons.
Homogeneous Subsets
The Homogeneous Subsets test was also done using Ryan-Einot-Gabriel-
Welsch Range based on the Mean Square Error of 67.585 and Harmonic
Mean Sample Size of 50.000 at .05 chosen alpha because the effect of the
independent variable (Types of Bulb) was statistically significant on the
dependent variable (Luminous Efficacy) to determine the means that fall
within the same homogeneous subset. Whenever the main effect of the
412
independent variable on the dependent variable is significant in ANOVA, it
confidently depicts that some group means or combination of group means
differ significantly from some other means or combination of group means.
The Ryan-Einot-Gabriel-Welsch Range test of Homogeneous Subsets is
suitably used for explicit specification of the means that fall inside one
subset and the means that fall inside another subset or other subsets. The
groups in the Ryan-Einot-Gabriel-Welsch Range table are shown in rows
that are ordered according to the magnitude of their means in ascending
order from the lowest to the highest mean; and according to columns that
represent subsets of means whose values significantly differ at the chosen
alpha level, .05 in this case. The means that are in one column are within
one subset, and are significantly different from the means that are in another
column or subset.
In the present example, there are three homogeneous subsets or columns,
each of which contains the mean of only one group or type of bulbs to
unquestionably show that each of the means is significantly different from
the other means. This table has shown that each of the three types of bulb
has a luminous efficacy mean that is significantly different from the others.
For instance, the Ryan-Einot-Gabriel-Welsch Range has revealed beyond all
possible statistical doubts that the 40W Incandescent bulbs have luminous
efficacy mean of 12.0000 that is overwhelmingly lower than the luminous
efficacy mean of 9W CFL bulbs (44.9800), which is in turn absolutely
lower than the luminous efficacy mean of the 8W LED bulbs (66.0000).
Each of the mean differences is statistically significant as no two means fall
under the same homogenous subset. And the Mean Difference in each case
is so large that it is significant statistically at as low as .000 Sig., read as
less than .0005 alpha.
413
whether it is a trivial, small, medium, or large effect size. This computation
is done by applying Cohen’s d. Cohen’s d for ANOVA Post Hoc Multiple
Comparisons to indicate Effect Size for each pairwise means difference is
equal to the Means Difference (Mean1 minus Mean2) divided by the square
root of the average Variances of the two groups. Average Variances of the
two groups represents the Std. Deviation1 squared plus the Std. Devaiation2
squared; divided by 2. Variance is the square of standard deviation. Put in
equation form, the Cohen’s d for ANOVA Pairwise Comparisons can be
as earlier given in Equation 6:4.
414
Recall from when the three types of t Test were treated that for
comparison of two means with Cohen’s d, the convention for trivial, small,
medium, and large Effect Sizes is as represented here in Table 6.2 for
instant reference. This classification corroborates the recommendations by
Cohen (1988).
415
0.500 to 0.799 Medium Effect Size
0.800 and above Large Effect Size
416
Exactly the same value of d is arrived at when the two equations are
used. So, when there is equal sample size in the groups under comparison,
this last equation that is easier to apply, should be used to get the effect size,
d in ANOVA Post Hoc Tests Multiple Comparisons of the dependent
variable with regard to the independent categorical variable.
For Pairwise Comparison of Means of the 9W CFL (M1 = 44.9800, SD1
= 8.65316) and 40W Incandescent (M2 = 12.000, SD2 = 5.03863), the
effect size, d is:
417
The Effect Size, d, of 4.658 for the difference between 9W CFL and
40W Incandescent bulbs is a very huge effect size; indicating that the
mean difference caused by the types of bulb on luminous efficacy is indeed
very awesome.
Again, the two samples are equal (N = 50 in each group), therefore the d
as a measure of Effect Size in ANOVA Post Hoc Tests Multiple Pairwise
Comparisons can be computed with great ease as:
418
Luminous Efficacy of the 9W CFL bulbs has a Mean (M1) of 44.9800
and a Standard Deviation (S1) of 8.65316. The Luminous Efficacy of the
40W Incandescent bulbs has a Mean (M2) of 12.0000 and a Standard
Deviation (S2) of 5.03863. Therefore, the Effect Size, d, is:
As can be seen, any of the two d equations used in the analysis, the
effect size remains exactly the same, 4.658. Use your discretion to apply
any of the two equations in analysing your research data for establishment
of Effect Size in ANOVA Post Hoc Tests Multiple Comparisons. But
always remember that having mastery of two approaches for execution of
any given statistical test, is typically more advantageous than knowing only
one method (Kpolovie, 2018; 2011). Try and master both techniques.
For Pairwise Comparison of the Means of the 8W LED (M1 = 66.000,
SD1 = 10.12372) and 40W Incandescent (M2 = 12.000, SD2 = 5.03863), d,
the Effect Size is:
419
The Effect Size, d, of 6.753 between the Luminous Efficacy Means of
8W LED and 40W Incandescent bulbs is an immensely large effect size.
The overwhelmingly large effect size (6.753) indicates that the luminous
efficacy mean difference between the two types of bulb is indeed a
profound difference that could only have been caused by the types of bulb
on luminous efficacy. Such massive difference could not have, in any way,
been caused by chance or some other factors.
Again, the two samples are equal in size (8W LED N = 50 and 40W
Incandescent N = 50). Therefore, the d as a measure of Effect Size in Post
420
Hoc Multiple Pairwise Comparisons of the luminous efficacy means of the
two types of bulb (8W LED and 40W Incandescent) can be computed with
great ease, using Equation 6:3.
Each of the two equations (Equation 6:4 and Equation 6:3) that was used
to compute the d, produced exactly the same Effect Size of 6.753 for the
difference between the luminous efficacy of 8W LED bulbs and 40W
Incandescent bulbs solely as a function of types of bulb. Use your discretion
to apply any of the two equations in analysing your research data whenever
you want to determine effect size for Post Hoc Tests Multiple Pairwise
Comparisons in Univariate ANOVA models. Mastering both of the
421
approaches has been my recommendation for students (Kpolovie, 2018;
2011).
422
used to light as many as five 8W LED bulbs. Similarly, the wattage
demanded to light one 40W Incandescent bulb can be used to light more
than four 9W CFL bulbs. Furthermore, while 40W Incandescent bulb has
1000 hours lifespan averagely, 9W CFL bulb has 10,000 hours lifespan on
the average, and the average lifespan of 8W LED bulb is 25,000 hours.
Therefore, the 8W LED bulb is indeed incomparably better than 9W CFL
bulb, which is in turn, incomparably better than the 40W Incandescent bulb.
In addition, while the 40W Incandescent bulb generates excessive heat as a
by-product, 8W LED bulb and 9W CFL bulb produce minimal heat.
An earlier study by the Pennsylvania State Centre for Nanoscale
Sciences arrived at similar findings when after displaying and observing
three types of bulbs in an investigation, stated that the “three bulbs in the
display each have a light output of 400 lumens, but they require different
amounts of power. The incandescent bulb uses 60 W, the fluorescent bulb
uses 7 W, and the LED bulb uses 6.5 W” (Penn State Center for Nanoscale
Sciences, 2016). Consequently, the use of high Watt incandescent light
bulbs has since 2012 been discontinued in the United States and in other
developed nations, but it is still the most popularly used bulb in many
Third-World countries, Nigeria in particular (Kpolovie and Ololube, 2020).
Haven completed the analysis and interpreted it, follow the remaining
routine protocols in IBM SPSS Statistics for data analysis to eventually shut
down your system.
View output
423
UNIVARIATE SPSS SYNTAX APPROACH FOR
ANOVA
The IBM SPSS Statistics Syntax for execution of the Univariate ANOVA
on effect of types of bulb on luminous efficacy is as given in Syntax 6.2.
Select Window
424
Select Go to Designated Syntax Window
Click Run
Click All
That is, start the computer, open IBM SPSS Statistics, and retrieve the
file that has been saved as Chapter 6 Table 1 Data.sav. The retrieval is
done by clicking Open data document icon in the Toolbar of SPSS
Data Editor to get its menu in which you look for and click the file name
(Chapter 6 Table 1 Data.sav) and press Enter (see Figure 6.32). Pressing
the Enter key on the Keyboard opens the Chapter 6 Table 1 Data.sav as
earlier shown in Figure 6.6.
425
Figure 6.32 Open Data menu box
426
Figure 6.34 IBM SPSS Statistics Syntax Editor
Figure 6.35 SPSS Syntax Editor with the Univariate ANOVA Syntax entered
Finally, click Run and click All (see Figure 6.36) to have the Output
displayed instantly as presented in Output 6.4.
427
Figure 6.36 Run All
428
429
430
431
432
INTERPRETATION OF OUTPUT 6.4
The Output 6.4 is exactly the same as Output 6.3. Therefore, the
interpretation of Output 6.3 is equally applicable for the Output 6.4.
Carefully attempt the interpretation of Output 6.4, using that of Output 6.3
as exemplary.
When done with the interpretation, the next thing to do is as follows.
View output
433
Save the output as Chapter 6 Output 4.spv and the Syntax as Chapter 6
Syntax 2.sps
434
Chapter 7
TWO-WAY ANALYSIS OF VARIANCE
Overview
When the possible effects of two independent variables on a dependent
variable are to be studied, Two-way ANOVA is the most appropriate
statistical test to adopt. It is a most powerful parametric inferential test for
simultaneous unquestionable establishment of two main effects and an
interaction effect of the categorical variables on the dependent variable.
Two-way ANOVA is a much more powerful statistical test than performing
double analysis of variance, one for each of the independent variables. With
the complete step by step screenshot illustrations, every user can
professionally apply two-way analysis of variance with SPSS syntax and
dialog boxes point and click. Effects of instructional techniques and partial
reinforcement schedules on students’ meta-mathematics achievement are
used for exemplification.
435
ANOVA quantitatively and qualitatively produces three effects, two main
effects and one interaction effect, with each of the effects representing a
separate partition of the variance of the dependent variable, as each of them
is separately tested for possible statistical significance. Each of the two
main effects in Two-Way ANOVA is a main effect of one of the two
independent variables on the dependent variable. The interaction effect is
the combined effect of the two independent variables on the dependent
variable (Vogt and Johnson, 2016; Landau and Everitt, 2004). In fact, the
interaction effect in Two-Way ANOVA is a higher order effect than the
main effects because it carries more information, and if the interaction
effect is significant statistically, it supersedes the main effects. The Two-
Way Analysis of Variance treated in this chapter is the Two-Way Between-
Subjects ANOVA.
Two-Way ANOVA demands two between-subjects independent
variables. The levels of the two independent variables are factorially
combined (combined in all possible ways) with each of the combinations
representing an independent group of cases. For instance, a Two-Way
ANOVA that one of the independent variables has three levels (categories)
and the other one has four levels, necessarily involves 12 independent
groups that are combined across the two variables as follows:
a1b1, a1b2, a1b3, a1b4;
a2b1, a2b2, a2b3, a2b4;
a3b1, a3b2, a3b3, a3b4.
This type of factorial design is commonly referred to as 3 x 4 (three by
four) Two-Way ANOVA. An example of the effects of Instructional
Techniques (with three levels) and Partial Reinforcement Schedules
(with four levels) on Meta-Mathematics achievement (the dependent
variable) will make this point more clearly.
436
i. Instructional techniques on students’ meta-mathematics achievement
ii. Partial reinforcement schedules on students’ meta-mathematics
achievement
iii. Interaction effect of instructional techniques and partial
reinforcement schedules on students’ meta-mathematics
achievement.
It engaged a small team of curriculum developers and psychologists for
the purpose. To accomplish the task, the team carried out an experiment on
the effects of instructional techniques and partial reinforcement schedules
on students’ meta-mathematics achievement. The team developed a meta-
mathematics curriculum content and delivered it to 60 randomly sampled
students who were further randomized into 12 groups of 5 each. The 12
groups of students were exposed to the following combination of treatment
conditions as in Table 7.0.
437
Table 7.1 Chapter 7 Table 1 Data.sav (Effects of InsTec and PRS on
MetMath)
Studs MetMath InsTec PRS
1 50 1 4
2 60 1 4
3 40 1 4
4 50 1 4
5 40 1 4
6 60 1 5
7 50 1 5
8 70 1 5
9 70 1 5
10 70 1 5
11 80 1 6
12 60 1 6
13 50 1 6
14 60 1 6
15 55 1 6
16 65 1 7
17 60 1 7
18 60 1 7
19 50 1 7
20 60 1 7
21 90 2 4
22 70 2 4
23 60 2 4
438
24 80 2 4
25 70 2 4
26 75 2 5
27 75 2 5
28 55 2 5
29 85 2 5
30 95 2 5
31 85 2 6
32 90 2 6
33 70 2 6
34 60 2 6
35 70 2 6
36 80 2 7
37 90 2 7
38 60 2 7
39 50 2 7
40 70 2 7
41 60 3 4
42 65 3 4
43 40 3 4
44 45 3 4
45 60 3 4
46 50 3 5
47 50 3 5
48 60 3 5
49 55 3 5
439
50 65 3 5
51 60 3 6
52 50 3 6
53 50 3 6
54 70 3 6
55 60 3 6
56 50 3 7
57 60 3 7
58 40 3 7
59 50 3 7
60 50 3 7
Research questions
i. What is the effect of instructional techniques on students’ MetaMaths
achievement as measured by their means and standard deviation?
ii. What is the effect of partial reinforcement schedules on students’
Meat-Maths achievement as measured by their means and standard
deviations?
iii. What is the interaction effect of instructional techniques and partial
reinforcement schedules on students MetaMaths achievement as
measured by their means and standard deviations?
Alternate hypotheses
i. Instructional techniques have significant effect on students’
MetaMaths achievement.
ii. Partial reinforcement schedules have significant effect on students’
MetaMaths achievement.
iii. Instructional techniques and partial reinforcement schedules have
significant interaction effect on students’ MetaMaths achievement.
Null hypotheses
440
i. Instructional techniques have no significant effect on students’
MetaMaths achievement.
ii. Partial reinforcement schedules have no significant effect on
students’ MetaMaths achievement.
iii. Instructional techniques and partial reinforcement schedules have no
significant interaction effect on students’ MetaMaths achievement.
Step 2. Open SPSS by clicking the icon of IBM SPSS Statistics version that
is installed in your computer. The SPSS icon looks like this, , or with
the name of the version like this, , depending on
the version installed in the computer. Allow it to fully get ready. Close the
Welcome to IBM SPSS Statistics dialog box that is superimposed on the
IBM SPSS Statistics Data Editor to have only the latter on your screen.
Step 3. Most carefully enter the data as they are shown in Table 7.1. The
entire data are to be entered in three columns just as was done for One-Way
ANOVA, just that the Two-Way ANOVA incudes a third column. The data
on the dependent variable (MetMath) are entered in the first column; the
data on the first independent variable (InsTec) which are the codes for the
three levels of Instructional Techniques (1 for Discovery, 2 for Inquiry
and 3 for Simulation) are entered in the second column; and data on the
second independent variable (PRS) that are the codes for the four levels of
Partial Reinforcement Schedules (4 for Fixed ratio, 5 for Variable ratio,
6 for Fixed interval and 7 for Variable interval) are entered in the third
column of the SPSS Data Editor.
After the careful data entry, the variables have to be named, given label
and provided values for the different levels of each of the independent
variables. Click Variable View at the bottom extreme left of the IBM SPSS
Statistics Data Editor, just beside Data View. This opens the Variable
441
View of the SPSS Data Editor for naming of the variables. In the first row
under Name column that is currently bearing VAR00001, type MetMath;
in the first cell under Label, type MetaMaths; and select Scale under
Measure.
In the second row under Name column that is currently bearing
VAR00002, type InsTec; in the cell of that row under Label, type
Instructional Techniques and under Measure, select Nominal. Then click
None under Values to see ellipsis (a small blue dotted box) , click the
ellipsis and Value Labels dialog box will appear. In it, type the code 1 in
the box beside Value and type Discovery in the box beside Label (see
Figure 7.1), and click Add.
Type the code 2 in the box beside Value, type Inquiry in the box beside
Label as shown in Figure 7.2, and click Add.
442
Figure 7.2 Value Labels 2 = Inquiry
Type the code 3 in the box beside Value, type Simulation in the box
beside Label, as in Figure 7.3, and click Add.
With this, the Value Labels for Instructional Techniques have been
completed as can be seen in Figure 7.4. Click OK that takes you back to the
starting point of the Variable View of the IMB SPSS Statistics Data
Editor.
443
Figure 7.4 Value Labels for Instructional Techniques completed
444
Figure 7.5 Value Labels 4 = Fixed ratio
Type the code 5 in the box beside Value, type Variable Ratio in the box
beside Label, exemplified in Figure 7.6, and click Add.
Type the code 6 in the Value box, type Fixed Interval in the Label box
as can be seen in Figure 7.7, and click Add.
445
Type the code 7 in the Value field, type Variable Interval in the Label
field as shown in Figure 7.8 and click Add.
With this, the Value Labels for the second independent variable (Partial
Reinforcement Schedules) is completed as presented in Figure 7.9.
Finally, click OK to get back to the starting point of the Variable View
of the IBM SPSS Statistics Data Editor that has been completed as
446
depicted in Figure 7.10.
Return to the Data View of the IBM SPSS Statistics Data Editor by
moving cursor to the bottom extreme left in the Variable View, and click
Data View. This shows the Data View of the SPSS Data Editor with all
the entered data and named variables as presented in Figure 7.11.
447
Figure 7.11 Chapter 7 Table 1 Data.sav
448
Figure 7.11 Continued
449
Figure 7.11 Continued
Step 4. Save the data by clicking the Save this document icon in the
Toolbar. In the box beside File name, type a suitable name for the data
document, Chapter 7 Table 1 Data.sav in this case, as can be seen in
Figure 7.12, and click Save or press Enter for the data to be securely saved
for subsequent retrieval and use whenever there is need for the dataset.
450
Figure 7.12 Save Data As… Chapter 7 Table 1 Data.sav
451
Figure 7.13 Analyze General Linear Model Univariate
452
Highlight and transfer MetaMaths from the left box into the Dependent
Variable box at the right by clicking the arrow that is facing the
Dependent Variable panel directly. Also highlight Instructional
Techniques and Partial Reinforcement Schedules in the left box and
move them into the Fixed Factor(s) box at the right by clicking the arrow
that is pointing at the Fixed Factor(s) as shown in Figure 7.15.
Click Plots pushbutton to have its dialog box called Univariate: Profile
Plots that is presented in Figure 7.16. In the Univariate: Profile Plots,
highlight PRS, the independent variable with more levels, and move it into
the Horizontal Axis field at the right by clicking the arrow pointing
directly at Horizontal Axis. Also, highlight InsTec, the independent
variable with fewer levels, and move it into the Separate Lines panel by
clicking the arrow directly facing the Separate Lines field.
453
Figure 7.16 Univariate: Profile Plots dialog box
Move cursor to and click Add for the two independent plots to actually
be added into the analysis in such a manner that will also allow for
interaction to be included. When they get added, PRS*InsTec appears in
the panel under Plots to show that the interaction effect of the two variables
will equally be provided in the Profile Plots as illustrated in Figure 7.17.
454
Figure 7.17 Univariate: Profile Plots with PRS*InsTec added
Click Continue to get back to the Univariate main dialog box. In it,
select Post Hoc pushbutton. With this, the Univariate: Post Hoc Multiple
Comparisons for Observed Means appears as indicated in Figure 7.18.
455
Figure 7.18 Univariate: Post Hoc Multiple Comparisons for Observed Means
456
Figure 7.19 Univariate: Post Hoc Multiple Comparisons with variables moved to Post Hoc
Test for
457
Figure 7.20 Univariate: Post Hoc Multiple Comparisons with R-E-G-W-Q Pairwise tests
checked
In the Univariate main dialog box this time, select Options pushbutton
for its dialog box to be displayed as in Figure 7.21.
458
Figure 7.21 Univariate: Options dialog box
459
Figure 7.22 Univariate: Options with checked Descriptive statistics and Estimates of effect
size
Click Continue to return to the main Univariate dialog box. From the
main Univariate dialog box, select EM Means pushbutton for Univariate:
Estimated Marginal Means dialog box to appear as shown in Figure 7.23.
460
Figure 7.23 Univariate: Estimated Marginal Means dialog box
461
Select Compare main effects to get its checkbox checked. Next,
click the small downward-pointed arrow and the names of some tests
(LSD, Bonferroni and Sidak) appear. Of the three tests, select Bonferroni
for the Confidence interval adjustment as shown in Figure 7.25.
Figure 7.25 Univariate: Estimate Marginal Means with Compare main effects checked and
Bonferroni chosen for Confidence interval adjustment
Then, click Continue to return to the Univariate main dialog box this
time when all the necessary statistical operations have been done (see
Figure 7.26).
462
Figure 7.26 Univariate main menu box when all the Two-Way ANOVA statistical
operations have been done
463
464
Estimated Marginal Means
465
466
467
468
Post Hoc Tests
469
INTERPRETATION OF OUTPUT 7.1 –
METAMATHS ACHIEVEMENT
The Two-Way ANOVA Output is very elaborate with 13 parts:
1. Univariate Analysis of Variance Between-Subjects Factors
2. Descriptive Statistics
3. Test of Between-Subjects effects
4. Estimated Marginal Means for Instructional Techniques
5. Pairwise Comparisons
6. Univariate Tests for Instructional Techniques
7. Estimated Marginal means for Partial Reinforcement Schedules
8. Pairwise Comparisons
470
9. Univariate Tests for Partial Reinforcement Schedules
10. Instructional Techniques*Partial Reinforcement Schedules
11. Post Hoc Tests for Instructional Techniques Homogeneous Subsets
12. Post Hoc Tests for Partial Reinforcement schedules
13. Profile Plots.
Each of these sections of the Output is very briefly interpreted.
Descriptive Statistics
This section of the Output has provided the Mean and Standard Deviation
of Meta-Mathematics Achievement for each of the Instructional Techniques
by Partial Reinforcement Schedules factorial combinations that serve as
answers to the research questions. It also shows the sample sizes (N) for the
various combinations of the independent variables. Recall that the example
in question is a 3 by 4 (3 x 4) factorial design that I had said will become
clearer as illustrated with this example. Perhaps, putting the means and
standard deviations in the cells of the 3 by 4 Two-Way ANOVA factorial
design as in Table 7.2 could clarify the descriptive statistics tabulated in the
second part of the Output7.1.
Table 7.2 Means and Standard Deviations for answering the research
questions
471
By putting the means and standard deviations into the 3 x 4 Two-Way
ANOVA factorial combinations has very clearly shown the mean and
standard deviation (in parenthesis) of the students’ MetaMaths achievement
in their respective cells across the two independent variables. The Marginal
Total (written in italics) mean and standard deviation for each of the three
levels of Instructional Techniques, and for the four levels of Partial
Reinforcement Schedules are also in the table. For the three levels of
Instructional techniques for instance, the Marginal Total mean and standard
deviation for Discovery are respectively 58.0000 and 10.18254, for Inquiry
are 74.0000 and 12.83499, and for Simulation are 54.5000 and 8.25578.
Similarly, for the four levels of Partial Reinforcement Schedules, the
Marginal Total mean and standard deviation respectively for Fixed ratio are
58.6667 and 14.93637, for Variable ratio are 65.6667 and 13.34523, for
Fixed interval are 64.6667 and 12.60197, and for Variable interval are
59.6667 and 12.88225. The Grand Mean and Standard Deviation, written in
bold italics, are 62.1667 and [13.47838], respectively.
472
Within-Subjects effect = Error) are the main concern in this Two-Way
ANOVA.
473
Thus, in the current example that the Sum of Squares InsTec is
4323.333 and the Sum of Squares Error is 5360.000, the Partial
Eta Squared (ɳp2) is:
474
The Effect Size, ɳp2 of .094 for the Partial reinforcement
schedules, a medium effect size, was arrived at by dividing the
Sum of Squares Partial Reinforcement Schedules (SSPRS) by the
Sum of Squares Partial Reinforcement Schedules (SSPRS) plus the
Sum of Squares Error (SSError). In this example that the SSPRS is
555.000 and the SSError is 5360.000, the Effect Size is
475
Though the interaction effect of Instructional techniques and
Partial reinforcement schedules on MetaMaths is not significant
statistically, its Partial Eta Squared of .082 is interpreted as a
medium effect size. This additional information accounts for why
it is good to indicate the effect size, irrespective of whether the
independent variable’s effect on the dependent variable is
significant or not.
476
This table has compared the Pairwise Means and shown that Discovery and
Inquiry have 16.000* Mean Difference that is statistically significant as the
.000 p-value (written as .001 or read as less than .0005) is lower than the
chosen .05 alpha in favour of Inquiry Instructional Technique, both Lower
and Upper Bounds (-24.290 and -7.710, respectively) falling below 0 (zero)
at 95% Confidence Interval of Difference. Discovery and Simulation have
3.500 Mean Difference, 3.342 Std. Error and .910 p-value that is not
significant, p > .05. Inquiry and Simulation have 19.500* Mean Difference
that is significant statistically, 3.342 Std. Error, p-value of .000 (read as less
than .0005) that is lower than .05 alpha. The Lower and Upper Bounds of
11.210 and 27.790, respectively, are both above 0 to further indicate the
overwhelming preponderant difference in favour of Inquiry Instructional
Technique. The other pairwise comparisons are interpreted in like manner;
and no other pair of the means differ significantly. Note that every Pairwise
Mean Difference that is significant has an asterisk. For all the Pairwise
Comparisons, Bonferroni Adjustment for multiple comparisons was used.
Univariate Tests
The Univariate Tests has further performed confirmatory F for the effect of
Instructional Techniques on students’ MetaMaths achievement on the basis
of the linearly independent pairwise comparisons among the estimated
marginal means. While the Contrast (i.e., the instructional techniques) has
4323.333 Sum of Squares, 2 df, and 2161.667 Mean Square; the Error
(Within-subjects differences) has 5360.000 Sum of Squares, 48 df and
111.667 Mean Square. The computed F is 19.358 that is statistically
significant even at p (Sig.) that is as low as .0005, and the Partial Eta
Squared that depicts Effect Size is as high as .446. Reject the null
hypothesis of no significant effect of Instructional Techniques on students’
MetaMaths achievement [F(2, 48) = 19.358, p < .05, ɳp2 = .446, a large
effect size]. This indicates that with all other factors held constant or under
complete control, only the Instructional Techniques accounts for at least
44.6% of the variance in students’ Meta Mathematics achievement.
477
This table has presented the Mean, Std. Error, and Lower and Upper
Bounds at 95% Confidence Interval for Fixed ratio, Variable ratio, Fixed
interval and Variable interval levels of Partial Reinforcement Schedules.
For instance, Fixed ratio has 58.667 Mean, 2.2728 Std. Error, 53.181 Lower
Bound and 64.153 Upper Bound at 95% Confidence Interval. Each of the
other three levels of Partial Reinforcement Schedules can simply be
interpreted in like manner.
Pairwise Comparisons
All the 12 possible Pairwise Comparisons on Partial Reinforcement
Schedules are presented in this table, showing the Mean Difference, Std.
Error, Sig., and the Lower and Upper Bounds at 95% Confidence Interval
for Difference based on estimated marginal means with Bonferroni
Adjustment for multiple comparisons. There is no Mean Difference that is
marked with asterisk to depict statistical significance. None of the 12
Pairwise Comparisons has a significant Mean Difference. That is, the
tabulated Mean Differences that range from 7.000 to 1.000 are better seen
as a function of chance like individual differences and measurement errors.
If the investigation is replicated 100 times with different samples (sized 60
each) drawn randomly from the same population of students, the results will
consistently indicate lack of significant difference in at least 95 times.
Univariate Tests
Special F tests were performed to further assess the effect of Partial
Reinforcement Schedules based on the linearly independent pairwise
comparisons among the estimated marginal means of MetaMaths
achievement. The one closest to reaching significant difference has 555.000
Contrast (Between-Groups, i.e. PRS) Sum of Squares, 5360.000 Error Sum
of Squares, 3 df for Contrast, 48 df for Error, 185.000 Contrast Mean
Square, 111.667 Error Mean Square, Sig. of .189, and F of 1.657 that is not
significant statistically. The Partial Eta Squared (Effect Size) is .094. Aptly,
the results of the effect of Partial Reinforcement Schedules on student’s
Meta-Mathematics achievement can technically be written as F(3, 48) =
1.657, p > .05, ɳp2 = .094, a medium effect size.
478
Instructional Techniques*Partial Reinforcement
Schedules
This table of the Output 7.1 has presented the Mean, Std. Error, Lower
Bound and Upper Bound at 95% Confidence Interval for the Interaction
effect of Instructional Techniques and Partial Reinforcement Schedules to
all the 12 possible factorial combinations. It could serve as part of the
answers to the research questions. For instance, Discovery combined with
Fixed Ratio has a Mean of 48.000, Std. Error of 4.726, Lower Bound of
38.498 and Upper Bound of 57.502 at 95% Confidence Interval.
479
of inquiry instructional technique on the average rather than discovery and
simulation instructional techniques.
Profile Plots
This graph is a pictorial illustration of the relative position of the Univariate
3 by 4 Two-Way ANOVA design Estimated Marginal Means of MetaMaths.
Discovery has marginal mean of 48.00 at Fixed ratio schedule, 64 at
Variable ratio schedule, 61 at Fixed interval schedule and 59 at Variable
interval schedule of reinforcement. Inquiry instructional technique has an
estimated marginal mean of 74 at Fixed ratio schedule, 77 at Variable ratio
schedule, 75 at Fixed interval schedule and 70 at Variable interval schedule
of reinforcement. The Inquiry instructional technique estimated marginal
means are very far above the means of other instructional techniques at all
levels of partial reinforcement schedules. Lastly, Simulation Instructional
Technique has estimated marginal mean of 54 at Fixed ratio schedule, 56 at
Variable ratio schedule, 58 at Fixed interval schedule and 50 at Variable
interval schedule of reinforcement. It is crystal clear from the graphic
illustration that the Inquiry Instructional Technique has overwhelmingly
positive effect on students’ MetaMaths achievement at each of the four
levels Partial Reinforcement Schedules with very high mean that ranges
from 70 and above. Each of the other two instructional techniques
480
(Simulation and Discovery) has MetaMaths achievement mean that is
smaller than 65 at each of the four levels of Partial Reinforcement
Schedule.
Haven completed interpretation of the findings, the Output 7.1, the next
thing to be done is adherence to the remaining six steps in the IBM SPSS
Statistics data analysis protocols as follows.
Step 6. You have already viewed the entire Output as the interpretation was
being done.
Step 7. Save the output by clicking the Save this document tool on the
Toolbar to have the Save Output As dialog box. In the box beside File
name, type a most suitable name for the document, Chapter 7 Output
1.spv, in this example and click Save as shown in Figure 7.27.
Step 8. Print the output. Move cursor to the Toolbar and click the Print tool
for its dialog box to appear. In the Print dialog box, confirm that All
481
visible output is checked (see Figure 7.28), and click Print for one copy of
the Output 7.1 to be printed. Repeat the action to get the required copies
printed.
Step 9. Exit IBM SPSS Statistics by closing all opened SPSS windows on
the system. When you click Close for the last Data Editor window on the
screen, SPSS will prompt you to confirm that you indeed want to exit the
software as shown in exhibit 7.29. Click Yes to proceed with exiting.
482
SPSS STATISTICS SYNTAX METHOD OF TWO-
WAY ANOVA
Application of IBM SPSS Statistics Syntax method is a very fast and
direct technique for execution of Two-Way Analysis of Variance. It simply
requires you to know the right Command Syntax and to correctly enter it
into the IBM SPSS Statistics Syntax Editor. You do not need to worry on
how and where to find the Syntax. The Syntax, like the ones for other
statistical tests, is provided readily in this book. The Syntax for execution of
Two-Way ANOVA is presented in Syntax 7.1.
This method for analysing the data requires starting the computer,
opening the IBM SPSS Statistics by clicking its icon on your computer. It
looks like this , or more fully with the name and the particular version,
looks like , depending on the version installed in
your system. When it fully comes up, close the superimposed dialog box
(Welcome to IBM SPSS Statistics) for you to actually activate the IBM
SPSS Statistics Data Editor.
483
In the SPSS Data Editor, retrieve the data file that was saved as
Chapter 7 Table 1 Data.sav by clicking File Open Data as
shown in Figure 7.30.
Your last action, a click on Data, instantly made the Open Data menu
box to appear as in Figure 7.31.
484
In the Open Data menu, look for the needed file, Chapter 7 Table 1
Data.sav, click it and press Enter key on the Keyboard. Once this is done,
the dataset file opens as displayed earlier in Figure 7.11.
Move cursor to and click Window to have its dropdown menu, and in
the Window dropdown, click Go to Designated Syntax Window,
illustrated in Figure 7.32.
Without any mistake, enter the Syntax provided in Syntax 7.1. As you
begin typing a word, SPSS will prompt you to make sure that no mistake is
made by producing some words that begin with the same alphabet(s) as the
one that you want to type. Just select the particular word that you have just
started typing. Selecting the words in this way will surely prevent mistakes
even if you have never personally typed any document before. When you
finish entering the Syntax, the IBM SPSS Statistics Syntax Editor will
look like the one demonstrated in Figure 7.34.
485
Figure 7.34 IBM SPSS Syntax Editor with the Syntax entered
Just in a split second, IBM SPSS Statistics completes the analysis and
produces the results in the IBM SPSS Viewer window as exhibited in
Output 7.2.
486
487
Estimated Marginal Means
488
489
490
491
Post Hoc Tests
492
493
INTERPRETATION OF OUTPUT 7.2
The Output 7.2 is equivalent to Output 7.1 and has the same interpretation
as the one provided for Output 7.1. Both Outputs were arrived at from
analysis of the same data in Table 7.1 that is saved as Chapter 7 Table 1
Data.sav. The only difference is that while IBM SPSS Statistics Syntax
method was applied on the data to produce Output 7.2, IBM SPSS Dialog
Boxes Point and Click method was employed in analysing the data that
resulted in Output 7.1. Carefully compare the Output 7.2 with Output 7.1 to
see if there are areas that they have different values.
After viewing the output, save it by clicking the Save this document
icon on the Toolbar. Then, type a suitable name for it (Chapter 7
494
Output 2.spv) in the box beside File name (see Figure 7.36) and press the
Enter key on the Keyboard.
Activate the Syntax Editor with the relevant Syntax in it and replicate
the actions for saving documents to save it as Chapter 7 Syntax 1.sps as
illustrated in Figure 7.37.
495
Figure 7.37 Save Syntax As… Chapter 7 Syntax 1.sps
Print the output and the syntax after which you close all opened SPSS
windows and exit the software. Click Start, click Power and click Shut
down.
496
Chapter 8
ONE-WAY REPEATED MEASURES
ANALYSIS OF VARIANCE
Overview
Nature is replete with recurrent events. Countless things occur repeatedly in
life. Majority of drugs, for instance, have to be taken several times for the
treatment of a disease. One-way repeated measures analysis of variance is
applied when each member of a sample received three or more measures on
the same dependent variable or on different related variables and the
concern is to ascertain whether observations taken after the treatments
yielded results that statistically differ significantly as the effects of the
experimental interventions, and nothing else, in the population. Effects of
nursing interventions on patients’ comprehensive health, and scholarly
progression in 5-year direct PhD engineering program are used for
demonstrations.
497
member of the sample received three or more measures on the same
dependent variable or on different variables and the concern is to ascertain
whether the different sets of scores for the same sample truly differ
significantly. In experimental design that the subjects are matched correctly
into groups after which the same measure is taken of them, the One-Way
Within-Subjects ANOVA could be used particularly when the scores
correlate very highly. One-Way Repeated Measures ANOVA is required for
analysis of data collected from longitudinal survey investigations that the
same measure is taken of the sample repeatedly over time (Kpolovie, 2017;
2016; 2010). In a One-Way Repeated Measures ANOVA, the means of an
independent variable are compared on a continuous dependent variable in
such a special manner that the independent variable is treated exclusively as
a within-subjects factor with three or more levels, and every of the
participants receives all the levels of the independent variable.
One-Way Repeated Measures ANOVA is employed as a most suitable
test of significant difference between means when multiple measures are
taken from the same group of participants. One-Way Within-Subjects
ANOVA breaks the sources of variation in the data into three: the variance
due to the difference in the means of the group (that is treated as between
groups mean square); the variance due to the fact that the measures are
correlated (such as individual uniqueness); and the variance due to the
residual or the remaining error. The Repeated Measures ANOVA
statistically puts the last two sources of variation together and treats it as the
residual and completely eliminates its influence on the difference due to the
means of the group’s different measures. This, One-Way Repeated
Measures ANOVA dose by dividing the between groups variance (the
between measures mean square) by the residual variance (error mean
square) to arrive at the F-ratio. When the different scores of the single
sample correlate substantially, the One-Way Between-Subjects ANOVA is
most efficient and remains the best possible test for establishment of
difference between the group’s means where such difference does exist. An
example will clarify the application of One-Way Within-Subjects ANOVA.
498
Suppose that the World Health Organization wanted to expand
breakthroughs in medicine and decided to engage the National Institute of
Nursing Research (NINR), Bethesda, Montgomery County, Maryland,
United States of America to conduct a research. And that the research was
aimed at determination of the Effects of Nursing Interventions on patients’
Comprehensive Health (total well-being). The National Institute of Nursing
Research decided to use Within-Subjects (Repeated Measures or
Correlated-Samples) Analysis of Variance design for the study. The NINR
structured three experimental treatment conditions, called Nursing
Interventions, for the investigation. The three Nursing Interventions are:
Physiological well-being
Social well-being
Psychological well-being.
Fifteen Nursing Doctors (5 for each intervention) were trained in
administration of Physiological well-being intervention, Social well-being
intervention, and Psychological well-being intervention conditions to the
participants. Execution of the experiment lasted for 210 days (seven
months). In the first month, measurements of Comprehensive Health were
taken and the average score for each of the participants was used as the
Baseline data. After that, all the 44 participants were exposed to the three
experimental treatment conditions (i.e., the Nursing Interventions) on
Physiological well-being, Social well-being, and Psychological well-being
during the same session in each of the remaining six months to control for
variations in their Comprehensive Health. To prevent possible or probable
influence of order (sequencing) effects and carry-over effects, the three
Nursing Interventions or treatment conditions were completely
counterbalanced in the administration. That is, taking Physiological well-
being intervention as A, Social well-being intervention as B, and
Psychological well-being intervention as C, these treatment conditions were
administered as exhibited in Table 8.0.
499
Apart from the Baseline data, the average of each participant’s six scores
on a given experimental intervention was taken as his or her score on that
particular treatment condition or intervention.
Further suppose that Biomedical Instruments were mainly used for the
measurement of Comprehensive Health of the 44 randomly sampled
participants (all aged 30) who had the same health conditions from the
onset. The Biomedical Instruments used are as enumerated.
Mechanical Waves – sound or ear waves and heart sounds.
Temperature – core, surface and ear.
Nucleic Acids – DNA and RNA.
Electrical Potentials – electrocardiogram (EKG) for heart,
electroencephalogram (EEG) for millivolt brain changes, and
electromyogram for muscle.
Gases – oxygen, carbon monoxide (CO) and carbon dioxide (CO2)
for lungs; arterial concentrations of oxygen (PaO2), carbon monoxide
(PaCO) and carbon dioxide (PaCO2).
Cellular – hormones, cytokines and cellular proteins with enzymes.
Pressure – veins (CVP), systolic and diastolic pressure plus MAP for
arteries; ICP for brain; uterus pressure monitor; esophagus (Eos);
lungs (Paw with intrapleural pressure).
Psychological well-being scale.
Social well-being scale.
To ensure that the various Biomedical Instruments plus the two self-
report scales yield comparable results, the measures obtained on each of the
instruments were transformed or standardized to T Score with a common
mean of 50 and a common standard deviation of 10. The average of each
participants’ T Score finally indicated his or her Comprehensive Health
500
after administration of each of the three Nursing Interventions monthly and
at the end of the experimental period as presented in Table 8.1.
501
22 10 20 10 20
23 90 90 90 90
24 70 80 70 70
25 80 80 70 60
26 40 30 50 50
27 90 90 90 100
28 40 60 50 50
29 60 70 70 60
30 50 60 60 50
31 50 70 50 50
32 60 70 70 60
33 50 60 60 50
34 20 40 40 30
35 50 70 60 50
36 50 70 50 60
37 40 50 40 50
38 70 80 70 70
39 40 50 50 40
40 30 50 30 30
41 60 70 60 60
42 30 40 40 40
43 80 90 80 80
44 60 80 60 60
Notation: *Pat. = Patients, Base. = Baseline, Phys. = Physiological, and
Psyc. = Psychological.
Research question
502
What is the effect of Nursing Interventions on patients’ Comprehensive
Health as measured by their means and standard deviations?
Alternate hypothesis
Nursing interventions have significant effect on patients’ Comprehensive
Health.
Null hypothesis
Nursing interventions do not have significant effect on patients’
Comprehensive Health.
Step 2. Launch IBM SPSS Statistics and close the superimposed Welcome
to IBM SPSS Statistics dialog box.
Step 3. Enter the data in the Data View of the IBM SPSS Data Editor
exactly as they are in Table 8.1. All the four scores of each participant must
constitute one row. The data entry for One-Way Repeated Measure is very
much like how data were entered for Paired-Samples t Test. The only
difference is that in One-Way Within-Subjects ANOVA, there are at least
three columns instead of the two columns in Paired-Samples t Test.
On completion of entering the data, click Variable View at the bottom
extreme left to have the Variable View of the IBM SPSS Statistics Data
Editor. In the Variable View of the Data Editor, type the actual names of
the four variables (each of the Nursing Interventions is treated as a
variable). Under Name column, type Baseline, Physiological, Social and
Psychological in place of VAR00001, VAR00002, VAR00003 and
VAR00004, respectively. Under Label, type Baseline well-being,
Physiological well-being, Social well-being and Psychological well-
being, respectively, in the four rows. Then select Scale for each of the
503
variables in the Measure column for the Variable View of the SPSS Data
Editor to look like Figure 8.1.
At the bottom extreme left of the SPSS Data Editor, click Data View to
return to the entered data when the variables have been named and labelled
as shown in Figure 8.2.
504
Figure 8.2 Chapter 8 Table 1 Data.sav
505
Figure 8.2 Continued
Step 4. Save the data by clicking the Save this document icon in the
Toolbar and type in a suitable name for the document in the box beside File
name as shown in Figure 8.3. In this example, Chapter 8 Table 1 Data.sav
is entered as the File name.
506
Figure 8.3 Save Data As Chapter 8 Table 1 Data.sav
507
Figure 8.5 Repeated Measures Define Factor(s) dialog box
In the Repeated Measures Define Factor(s) dialog box, type the name
of the Within-Subjects Factor in the box beside Within-Subject Factor
Name, which in this example is Interventions. The Factor Name should
ideally be a single word or phrase without spacing and that is why
Interventions is entered instead of Nursing Interventions. Also, in this
dialog box, provide the number of levels in the Within-Subject Factor in
the small box beside Number of Levels. Type the number, 4, in the small
box as shown in Figure 8. 6. In addition, type Comprehensive_Health, the
name of the dependent variable, in the field for Measure Name. Again, the
dependent variable name must be a single word or phrase without space.
So, underscore is used to make Comprehensive Health as a single word.
508
Figure 8.6 Repeated Measures Define Factor(s) with the Factor Name, Number of Levels
and Measure Name
509
Figure 8.7 Repeated Measures Define Factor(s) with name and level added
Click Define for the main Repeated Measures dialog box to appear as
in Figure 8.8 for further statistical operations to be done. Most of the
required statistical operations are done in the Repeated Measures dialog
box.
510
Figure 8.8 Repeated Measures dialog box
Move each of the variables from the left box into the Within-Subjects
Variables (Interventions) box at the right by highlighting it and clicking
the right-pointed arrow according to the numbered order. When done,
the dialog box will look as shown in Figure 8.9. You can also move all of
the variables at once by pressing and holding down the Ctrl key on the
Keyboard and selecting all the four variables and then clicking the right-
pointed arrow that is directly facing the Within-Subjects Variables
(Interventions) panel.
511
Figure 8.9 Repeated Measures dialog box with levels of the Factor moved to the right
Select the Options pushbutton for its dialog box to appear. The
Repeated Measures: Options dialog box, shown in Figure 8.10, has a list
of several possible statistical operations that could be opted for, depending
on types and volume of information you want in the specific statistical test
(One-Way Within-Subjects ANOVA in this case) that you are executing.
512
Figure 8.10 Repeated Measures Options
513
Figure 8.11 Repeated Measures: Options with Descriptive statistics and Estimate of effect
size checked
514
Figure 8.12 Repeated Measures: Estimated Marginal Means
515
Figure 8.13 Repeated Measures: Estimated Marginal Means with Interventions moved
under Display Means for
516
Figure 8.14 Repeated Measures: Estimated Marginal Means with Compare main effects
and Bonferroni selected
Click Continue to get back to the main menu for Repeated Measures.
Select Plots pushbutton for the Repeated Measures: Profile Plots dialog
box to appear as shown in Figure 8.15.
517
Figure 8.15 Repeated Measures: Profile Plots
518
appears in the Horizontal Axis field. Then, click Add. On clicking Add,
the Interventions disappears from the Horizontal Axis box and appears in
the big panel under Plots. Select Bar Chart and check Include reference
line for grand mean as shown in Figure 8.16.
Figure 8.16 Repeated Measures: Profile Plots with Interventions added, Bar Chart and
Include reference line for grand mean checked
519
Click Continue to return to Repeated Measures main dialog box when
all the necessary statistical operations have been completed. Finally, click
OK for IBM SPSS Statistics to complete running the analysis and
displaying the results in the IBM SPSS Viewer window as shown in Output
8.1.
520
Figure 8.17 Paired-Samples T Test
Figure 8.18 Paired-Samples T Test as follow-up for the One-Way Repeated Measures
ANOVA according to the required pairwise comparisons
521
Finally, move cursor to and click OK for IBM SPSS Statistics to
complete the analysis and present the results in the IBM SPSS Statistics
Viewer window as exhibited in the further follow-up pairwise comparisons
part of the Output 8.1.
522
523
524
525
Further Pairwise Tests using Paired-Samples t Test
526
527
INTERPRETATION OF OUTPUT 8.1: EFFECTS OF
NURSING INTERVENTIONS
The One-Way Repeated Measures Analysis of variance results presented in
Output 8.1 has 11 parts. They are the Within-Subjects Factor, Descriptive
Statistics, Multivariate Tests, Mauchly’s Test of Sphericity, Tests of Within-
Subjects Effects, Tests of Within-Subjects Contrasts, Tests of Between-
Subjects Effects, Estimated Marginal Means Interventions, Pairwise
Comparisons, Multivariate Tests, and Profile Plots. In addition to these,
there are three sections (Paired Samples Statistics, Paired Samples
Correlations and Paired Samples Test) in the Further Pairwise Comparisons
done with Correlated t Test as a follow-up analysis.
528
Within-Subjects Factors
This part of the Output simply lists the four Nursing Interventions (Baseline
coded 1, Physiological coded 2, Social coded 3, and Psychological coded 4)
to show that the means compared were got from the dependent variable
(Comprehensive Health) scores on each of these four levels of the
independent variable, Nursing Interventions).
Descriptive Statistics
This section is a display of the Mean, Std. Deviation and N for each of the
four Interventions levels. The N is 44 across board as only one sample of 44
participants was studied and every of the participants experienced the
Baseline observations, as well as the monthly observations after exposure to
the experimental treatment conditions. For instance, Baseline well-being
has 49.7727 Mean and 20.05674 Std. Deviation, Physiological well-being
has 62.2727 Mean and 18.28249 Std. Deviation, Social well-being has
55.0000 Mean and 18.36326 Std. Deviation, and for Psychological well-
being, the Mean is 52.0455 and the Std. Deviation is 18.11841. These
Descriptive Statistics information serve as the answer to the posed research
question.
Multivariate Tests
This section of the Output shows that the null hypothesis of ‘no significant
effect of nursing interventions on participants’ comprehensive health’ can
be tested either with univariate test (One-Way Within-Subjects ANOVA) or
with a Multivariate test (One-Way Within-Subjects Multivariate Analysis of
Variance [MANOVA]). The Multivariate Tests presented results for four
different tests (Pillai’s Trace, Wilks’ Lambda, Hotelling’s Trace, and Roy’s
Largest Root) (Hair Jr., Black and Anderson, 2010; Johnson and Wichern,
2014) on the effect of the nursing Interventions on patients’ comprehensive
health. The F (44.723) for each of these tests is statistically significant at
even .001 alpha as the Sig. is .000 (i.e., less than .001, also written as less
than .0005); and with as high as .766 Partial Eta Squared (the Effect Size).
However, it is not this Multivariate tests results that are the concern of the
example; the MANOVA results merely happen to be part of the output of
529
the One-Way Repeated Measures ANOVA that is the issue under
consideration.
530
Interventions between measures divided by the Mean Square for
Interventions error on the basis of any of the four tests (Sphericity
Assumed, Greenhouse- Geisser, Huynh-Feldt or Lower-bound). In equation
form F is arrived at thus in Equation 8:1.
Substituting the numerator and denominator with their actual values for
Sphericity Assumed, F is:
Substituting the numerator and denominator with their actual values for
Greenhouse-Geisser (that is when Sphericity is not assumed), F is:
531
chosen alpha of .05 and consequently, the null hypothesis of “Nursing
interventions have no significant effect on patients’ Comprehensive Health”
is rejected. In other words, Comprehensive Health significantly differ
across the four levels of types of Nursing Interventions (Baseline well-
being, Physiological well-being, Social well-being, and Psychological well-
being) in the population. That is, the Comprehensive Health means of the
four repeated measures, based on the Nursing Interventions, significantly
differ. The experimental intervention conditions directly and singly account
for as much as 49.1% of the variance in Comprehensive Health in the
population as the Effect Size (Partial Eta Square) is .491.
Effect Size
The Effect Size that is computed with Partial Eta Squared (ɳp2) is equal
to the Type III Sum of Squares for Interventions (or Effect) divided by the
Type III Sum of Squares for Intervention plus Type III Sum of Squares for
Error. In equation form, Partial Eta Squared is as in Equation 8:2.
532
Interventions. The Means shown here are the same as provided by the
Descriptive Statistics. In the Estimated Marginal Means, Std. Error is
provided unlike in the Descriptive Statistics section where Std. Deviation
instead was given for each of the Interventions. For instance, the fourth
nursing intervention (Psychological well-being intervention) has a mean of
52.045, Std. Error of 2.731, Lower Bound of 46.537 and Upper Bound of
57.554 at 95% Confidence Interval. Each row in the Estimated Marginal
Means for Interventions is interpreted in like manner.
Pairwise Comparisons
The Pairwise Comparisons table has presented the Mean Difference (I-J),
Std. Error, Sig., and 95% Confidence Interval for Difference Lower Bound
and Upper Bound for each of the Pairwise Comparisons. Under the Mean
Difference (I-J) column, each Mean Difference that is statistically
significant has an asterisk. To show the direction that the Mean Difference
favours, where the mean of the (I) Interventions is smaller than the mean of
the (J) Interventions, a minus sign (-) is put as the prefix of the given Mean
Difference because the (J) mean is subtracted from the (I) mean to arrive at
the Mean Difference. Where the mean of the (I) Interventions is greater
than the mean of the (J) Interventions, there is no minus sign as prefix of
the Mean Difference.
`The participants’ Comprehensive Health mean at the Baseline (coded 1)
is significantly lower than the Comprehensive Health mean at Physiological
well-being Intervention (coded 2) with [p < .001], and at Social well-being
(coded 3) with [p < .001]; but does not differ significantly from the
Comprehensive Health mean at the Psychological well-being (coded 4) with
[p > .05]. The Comprehensive Health mean at Physiological well-being
Intervention is significantly greater than the Comprehensive Health mean at
each of the other three Interventions (Baseline well-being, Social well-
being and Psychological well-being) with p < .001 in each case. Social
well-being has Comprehensive Health mean that is significantly greater
than that at the Baseline (p < .001), and significantly smaller than that at
Physiological well-being Intervention; but does not significantly differ from
that at Psychological well-being Intervention. Lastly, the Comprehensive
Health mean at Psychological well-being Intervention does not differ
significantly from that at the Baseline well-being Intervention and Social
533
well-being Intervention; but is significantly smaller than the
Comprehensive Health mean at Physiological well-being Intervention.
Of the four Nursing Interventions, the Physiological well-being
Intervention is the most effective with overwhelming preponderance,
followed by the Social well-being Intervention. The fourth well-being
Intervention, Psychological, does not differ from the Baseline well-being
Intervention; and is significantly lower than the Physiological well-being
Intervention.
Multivariate Tests
The Multivariate Tests results here are exactly as those reported under the
third section of the Output 8.1. Each of the F for the four tests (44.723) is
statistically significant with p < .001, and has .766 Partial Eta Squared
(Effect Size). Each of the F tests the multivariate effect of Nursing
Interventions on the basis of linearly independent pairwise comparisons
among the estimated marginal means.
Profile Plots
This has graphically illustrated the magnitude and relative position of each
Nursing Intervention’s effect on Comprehensive Health as shown pictorially
by the Mean as well as the extent to which each of the Means differs from
the Observed Grand Mean (the horizontal dark line). Recall that scores
yielded by each of the various measuring instruments were standardized to
T Score with a mean of 50 and a standard deviation of 10. This accounts for
why the Profile Plots has shown some sort of closeness of the
Comprehensive Health mean for each Nursing Intervention to 50. The Bar
Chart bar for Baseline well-being is approximately 50, Physiological well-
being Intervention is approximately 62, Social well-being Intervention is
55, and Psychological well-being Intervention is approximately 52. The
Observed Grand Mean which is the mean for all the Nursing Interventions
put together is approximately 55.
534
Deviation and Std. Error Mean for each Intervention in the six Pairwise
comparisons. The statistical information is the same with those shown
under Descriptive Statistics section of the General Linear Model Output.
535
than .0083, the null hypothesis of no significant mean difference between
the two Interventions will be retained.
Simply put, .0083 is to be used as the chosen alpha in testing the
tenability of each of the six pairwise comparisons. This is to ensure that the
probability of committing Type I error (false rejection of a null hypothesis
that should by right have been retained) does not become greater than .05
for the entire set of the follow-up tests. The implication is that if no
adjustment was made and each of the six follow-ups pairwise tests was
conducted at .05, the experimental eta, the overall alpha for the set of tests
would have been increased to the sum of the .05 alphas for the six tests,
which is .300.
The Paired Samples Test of Paired Differences has shown the Mean
Difference, the Difference in Std. Deviation, Std. Error Mean, and the
Lower and Upper 95% Confidence Interval of Difference, the t, df and Sig.
(2-tailed) for each of the six pairwise comparisons thus:
1. Baseline well-being and Physiological well-being interventions has a
Mean Difference of -12.500, t of -11.041 that is significant
statistically as p < .0083.
2. Baseline well-being and Social well-being interventions has a Mean
Difference of -5.227, t of -5.518 that is significant statistically as p <
.0083.
3. Baseline well-being and Psychological well-being interventions has a
Mean Difference of -2.273, t of -2.226 that is not significant
statistically as p > .0083.
4. Physiological well-being and Social well-being interventions has a
Mean Difference of 7.273, t of 5.220 that is significant statistically
because p < .0083.
5. Physiological well-being and Psychological well-being interventions
has a Mean Difference of 10.227, t of 7.125 that is significant
statistically because p < .0083.
6. Social well-being and Psychological well-being interventions has a
Mean Difference of 2.955, t of 2.562 that is not significant
statistically because p > .0083.
These six pairwise results are exactly like the results arrived at with the
Pairwise Comparisons in section 9 of the Output 8.1. That is, the pairwise
comparisons that have significant difference in the Paired Samples Test of
536
Paired Differences are exactly the ones that were significantly different in
the General Linear Model Pairwise Comparisons.
Step 6. View the entire output scrolling up and down, right and left.
Step 7. Save the output. Click the Save this document icon to have its
dialog box. Type the name, Chapter 8 Output 1.spv, for the document in
the box beside File name (see Figure 8.19) and click Save.
Step 8. Print the output after which you print the data. Follow the print
command to get this done. You simply click the Print tool in the
Toolbar to have its dialog box in which you click Ok when the All visible
output is checked (see Figure 8.20) for one copy of All visible output to be
printed.
537
Figure 8.20 Print dialog box
Step 9. Exit SPSS by closing all opened IBM SPSS Statistics windows.
Step 10. Shut down the system be clicking Start, followed respectively by
clicking Power and clicking Shut down.
538
Boot the computer. Launch IBM SPSS Statistics by clicking its icon
. When it fully comes up, close the superimposed
Welcome to IBM SPSS Statistics dialog box. Move cursor to the Toolbar
and click the Open data document tool to have its menu as given in
Figure 8.21. In the Open Data menu, look for and click the name with
which the data document was saved (Chapter 8 Table 1 Data.sav) and
press Enter key on the Keyboard.
539
Figure 8.21 Open Data
With the pressing of Enter, the document opens exactly as much earlier
presented in Figure 8.2. Click Window and from its dropdown menu, click
Go to Designated Syntax Window (see Figure 8.22).
540
Statistics Syntax Editor can also be opened by clicking File New
Syntax.
Most carefully enter the Syntax that is provided in Syntax 8.1. When
done, it will look like Figure 8.24.
Figure 8.24 IBM SPSS Statistics Syntax Editor with the requisite syntax
541
Output 8.2 Chapter 8 Output 2.spv
542
543
544
545
Further Pairwise Tests with Paired-Samples t Test
546
547
INTERPRETATION OF OUTPUT 8.2
Compare every table in Output 8.2 with its corresponding table in Output
8.1 and list the differences. At the end, how many differences were you able
to spot and list? Zero difference. Right, if there is no difference at all in the
two outputs, then the interpretation provided for Output 8.1 cannot be
different from how Output 8.2 should be interpreted. Apply the
interpretation of Output 8.1 to the Output 8.2.
Follow the remaining IBM SPSS Statistics protocols to eventually shut
down your system.
View output
548
Save the output as Chapter 8 Output 2.spv and the Syntax as Chapter 8
Syntax 1.sps
549
Indian Institute of Science is one of the best universities in the world to do a
Ph.D. in Engineering. There are several options for it but this example is on
the 5-Year Direct program alone.
Imagine that a researcher wanted to ascertain his conjecture that the
scholarly performance of the admitted students in Indian Institute of
Science significantly improved over the 5-year duration of the Ph.D.
program in engineering. He was particularly concerned with whether the
students’ scholarly performance significantly changed positively or
negatively across the five years. Suppose that the students’ annual
performance was in scale data that were transformed into T Score as
presented in Table 8.2. It was suspected that the students’ scholarly
performance would either improve over the years as they become more
acclimated to the technicalities of their subject matter and
instructors/supervisors’ approach; or decrease as they become overwhelmed
by the increased complexity of the subject matter and fieldwork.
550
551
Research question
How is the students’ scholarly performance across the 5-year Direct Ph.D.
engineering program as measured by their means and standard deviations?
Alternate Hypothesis
552
The students significantly differ in their scholarly performance over the 5-
year Direct Ph.D. engineering program.
Null Hypothesis
The students do not differ significantly in their scholarly performance over
the 5-year Direct Ph.D. engineering program.
Step 3. Enter the data exactly the way they are in Table 8.2, just the same
way that Table 8.1 data were typed in. Make sure that each person’s five
scores are consistently in one row without any mistake. When done, click
Variable View and type in the actual names, Year1SP, Year2SP, Year3SP,
Year4SP and Year5SP respectively in place of Var00001, Var00002,
Var00003, Var00004 and Var00005 rows under Name column. Type in the
same five names for the variables in the Label column. On completion, the
Variable View of the IBM SPSS Statistics Data Editor will look like
Figure 8.25.
Figure 8.25 Completed Variable View for the One-Way Within-Subjects ANOVA
553
Move the cursor to and click Data View to have the Data View of the
SPSS Data Editor with the entire data and names for the variables
(Scholarly Performance [SP] for each year is treated as a variable) as
presented in Figure 8.26.
Figure 8.26 Data Editor with the data and named variables
554
Figure 8.26 Continued
555
Figure 8.26 Continued
Step 4. Save the data by clicking the Save this document tool in the
Toolbar and typing the document name, Chapter 8 Table 2 Data.sav in the
File name box, and clicking Save (see Figure 8.27).
556
Figure 8.27 Save Data As… Chapter 8 Table 2 Data.sav
557
With the last click, Repeated Measures Define Factor(s) dialog box
appeared as shown in Figure 8.29.
558
Figure 8.30 Repeated Measures Define Factor(s) with the names and number of levels
entered
Click Add, for both the Levels and Name so that the Factor and its
Levels as well as the Measure Name (the dependent variable,
Performance as short form for Scholarly Performance) will be added into
the analysis. When added, the names for the Factor Name and Measure
Name will respectively appear in the upper and lower big boxes beside
as shown in Figure 8.31.
559
Figure 8.31 Repeated Measures Define Factor(s) when the Factor (Year) and Measure
(Performance) have been added
Click Define so that the actual Repeated Measures main dialog box
will appear as given in Figure 8.32.
560
Figure 8.32 Repeated Measures dialog box
561
Figure 8.33 Repeated Measures dialog box with the factor levels moved to the right
562
Figure 8.34 Repeated Measures: Profile Plots dialog box
The Year under Factors box is highlighted. So, click the right-pointed
arrow next to it in order to move it into the Horizontal Axis box at the
right as demonstrated in Figure 8.35.
563
Figure 8.35 Repeated Measures: Profile Plots with Year moved under Horizontal Axis
Click Add, so that the Profile Plots can be completed. After that click,
the Year will appear in the big box just under Plots instead of remaining in
the Horizontal Axis field. Next, select Bar Chart, and check the checkbox
for Include reference line for grand mean so that the relative variation of
each of the Factor Means from the Grand Mean can be visualized in the
results (see Figure 8.36).
564
Figure 8.36 Repeated Measures: Profile Plots with Year added, Bar Chart and Include
reference line for grand mean checked
565
Figure 8.37 Repeated Measures: Estimated Marginal Means dialog box
Figure 8.38 Repeated Measures: Estimated Marginal Means with Year moved
Check the Compare main effects checkbox. Then click the small black
arrow that is pointing downwards to have its dropdown menu in which
566
you click Bonferroni for the Confidence interval adjustment as can be
seen in Figure 8.39.
Figure 8.39 Repeated Measures: Estimated Marginal Means with Compare main effects
checked and Bonferroni selected for Confidence interval adjustment
Click Continue to get back to the Repeated Measures main dialog box.
In it this time, select Options. This produces Repeated Measures: Options
dialog box as depicted in Figure 8.40.
567
In the Repeated Measures: Options dialog box, check the small boxes
beside Descriptive statistics and Estimates of effect size for the results of
these analysis to be part of the eventual output (see Figure 8.41). The
Estimates of effect size performs the much-needed Partial Eta Squared to
automatically show the Effect Size of the independent variable (Year of
study) on the dependent variable Scholarly Performance) of students in
Indian Institute of Science 5-year Direct Ph.D. program.
Figure 8.41 Repeated Measures: Options with Descriptive statistics and Estimates of effect
size checked.
568
Figure 8.42 Analyze Compare Means Paired-Samples T Test
569
Figure 8.43 Paired-Samples T Tests as follow-up pairwise comparisons for the One-Way
Repeated Measures ANOVA
570
571
572
573
Pairwise Tests as Follow-up
574
575
576
INTERPRETATION OF THE OUTPUT 8.3 –
INFLUENCE OF YEARS OF STUDY
There are 14 sections in the robust output that IBM SPSS Statistics has
produced for the One-Way Repeated Measures Analysis of Variance as
briefly examined herein. While the first 11 parts of the output are purely for
the One-Way Within-Subjects ANOVA, the additional three sections are for
Paired-Samples t Test as a follow-up analysis.
Within-Subjects Factors
The five years that observation of the Scholarly Performance was done are
listed in this table, and accordingly coded as 1, 2, 3, 4 and 5, respectively.
Descriptive Statistics
This table has given descriptive statistics (Mean, Std. Deviation and N) of
the Scholarly Performance for each of the five years. The N (number of
students or scholars) for each of the years is 68. The Mean and Std.
Deviation in Year1 are respectively 59.1765 and 9.10624; in Year2 are
64.9559 and 9.18375, respectively; in Year3 are 75.3971 and 8.77435,
respectively; in Year4 are respectively 80.8676 and 9.36316; and in Year5
are 90.7206 and 9.91583, respectively. There is an incremental trend in the
means over the years.
Multivariate Tests
This table shows that the null hypothesis that “the students do not differ
significantly in their scholarly performance over the 5-year Direct Ph.D.
engineering program” can be tested either with univariate test (One-Way
Within-Subjects Analysis of Variance [ANOVA]) or with a multivariate test
(One-Way Within-Subjects Multivariate Analysis of Variance [MANOVA]).
With the Multivariate Tests, results for the four different tests (Pillai’s
Trace, Wilk’s Lambda, Hotelling’s Trace, and Roy’s Largest Root) on the
influence of years of study on the students’ scholarly performance is
statistically significant in each case (in each of the four tests) as the F is
1453.813 with df Hypothesis of 4.000 and df Error of 64.000; and Sig. of
.000 that is less than .0005; plus a very high Effect Size measured with
577
Partial Eta Squared to be .989. However, it is not the Multivariate tests
results that are the overall concern of this example. The MANOVA results
merely happen to be part of the comprehensive output produced by SPSS
for the One-Way Repeated Measures ANOVA that is the main issue under
consideration in this example.
578
Tests of Within-Subjects Effects
The most critical and completely indispensable to report and interpret of all
the sections in the One-Way Repeated Measures ANOVA output is the
part that is called Tests of Within-Subjects Effects for testing the tenability
of the postulated null hypothesis. The Tests of Within-Subjects Effects
provides the much-needed F test as the ratio of the two sources of variance,
each represented as a Mean Square (MS), that make up the General Linear
Model Univariate Analysis of Variance whenever repeated measures are
taken. The F is the Mean Square for Effect (Year between measures)
divided by the Mean Square for Error (Residual) across the years for each
of the four F tests (Sphericity Assumed, Greenhouse-Geisser, Huynh-Feldt
and Lower-bound). In equation form, F is arrived at as in Equation 8:1.
For the Greenhouse-Geisser (as the sphericity is not assumed), when the
numerator and denominator are substituted with their actual values, F is:
For Sphericity Assumed (if the sphericity were actually assumed), when
the numerator and denominator are substituted, F is:
579
The Tests of Within-Subjects Effects has provided the Type III Sum of
Squares (SS), df, Mean Square, F, Sig. and Partial Eta Squared (Effect
Size) for each of the four statistical test (Sphericity Assumed, Greenhouse-
Geisser, Huynh-Feldt and Lower-bound) that are due to variance between
the different years’ scholarly performance on the one hand, and due to the
residual or error (variance in each year’s scholarly performance) of the
students in the Direct 5-year Ph.D. Engineering program in the Indian
Institute of Science.
Taking Greenhouse-Geisser, as the Sphericity Assumption is not met,
the Type III Sum of Squares for Effect is 42838.482 and for Error is
979.118; df for Effect is 2.622 and for Error or Residual is 175.643; Mean
Square for Effect is 16340.988 and for Error is 5.574; the F is 2931.393; the
Sig. is .000 (read as less than .0005); and the Partial Eta Squared is .978.
The computed F of 2931.393 is statistically significant because the Sig.
(.0005) is lower than the chosen alpha of .05; therefore the null hypothesis
that “the students do not differ significantly in their scholarly performance
over the 5-year Direct Ph.D. engineering program” is rejected; and the
Effect Size measured rightly with Partial Eta Squared (ɳp2) is .978.
Rejection of the null hypothesis is because [F(2.622, 175.643) = 2931.393,
p < .05, ɳp2 = .978, a profoundly large effect size]. The students in the
program significantly differ in their scholarly performance over the 5-year
duration of the program. Pairwise Comparisons to be examined shortly
from now will unveil the nature, pattern or direction of the difference.
Based on Sphericity Assumed (in case the Sphericity Assumption were
met), Type III Sum of Squares for Effect is 42838.482 and for Error is
979.118; df for Effect is 4 and for Error is 268; Mean Square for Effect is
10709.621 and for Error is 3.653; F is 2931.393; Sig. is .000 (read as less
than .0005); and Partial Eta Squared (Effect Size) is .978. Since the Sig. of
.0005 is lower than the classically chosen alpha of .05, the null hypothesis
that “the students do not differ significantly in their scholarly performance
over the 5-year Direct Ph.D. engineering program” is rejected; and the
Effect Size as measured by Partial Eta Squared is .978. The F of 2931.393
is statistically significant because the null hypothesis is rejected for the
reason that [F(4, 268) = 2931.393, p < .05, ɳp2 = .978, a profoundly large
effect size].
In both Sphericity Assumed and Sphericity not Assumed, the Sig. is less
than the chosen alpha of .05 and even of .01 (if this were to be the chosen
580
level of significance). Consequently, “the null hypothesis that the students
do not differ significantly in their scholarly performance over the 5-year
Direct Ph.D. engineering program” is rejected. In other words, the students’
scholarly performance significantly differs across the 5-year duration of the
Direct Ph.D. engineering program in the Indian Institute of Science. That is,
the scholarly performance means in the 5 years duration of the program are
significantly different, denoting that the 5 years have statistical effect on the
program, accounting for as much as 97.8% of the variance in scholarly
performance, as the Effect Size (Partial Eta Squared) is .978.
Effect Size
The Effect Size that is computed with Partial Eta Squared is derived by
dividing the Type III Sum of Squares for Effect (SSEffect) of Year by Type
III Sum of Squares for Effect (SSEffect) of Year plus the Type III Sum of
Squares for Error (SSError)of a given F test. In the form of equation, Partial
Eta Squared for Greenhouse-Geisser F test is as given in Equation 8:2.
The Partial Eta Squared for Sphericity Assumed F test is still as in the
Equation 8:2.
581
Test of Between-Subjects Effects and Test of Wuthin-
Subjects Contrast
These two tests (Test of Within-Subjects Contrast and Tests of Between-
Subjects Effects) are not required in the reporting and interpretation of One-
Way Repeated Measures ANOVA. They may be required in some other
statistical tests where they necessarily play important roles.
Pairwise Comparisons
The Pairwise Comparisons section of the Output 8.3 is crucially essential in
reporting the results. It has presented the Scholarly Performance Mean
Difference (I-J), Std. Error, Sig., and 95% Confidence Interval for
Difference Lower Bound and Upper Bound for each of the Pairwise
Comparisons across the Years. Under the Mean Difference (I-J) column,
each Mean Difference that is statistically significant has an asterisk (*) and
the Sig. for each pair of Means is less than .0005 (p < .0005). The asterisk
runs across all the Pairwise Comparisons in favour of the higher Year.
Therefore, for each of the Comparisons, the Mean Difference is statistically
significant with each of the higher Years having greater Mean.
To show the direction that the Mean Difference favours, where the mean
of the (I) Year is smaller than the mean of the (J) Year, a minus sign (-) is
put as the prefix of the given Mean Difference because the (J) mean is
subtracted from the (I) mean to arrive at the Mean Difference. Where the
mean of the (I) Year is greater than the mean of the (J) Year, there is no
minus sign as prefix of the Mean Difference. The evidence of statistical
582
significance in each Mean Difference can also be seen in the 95%
Confidence Interval for Difference Lower Bound and Upper Bound because
wherever the Mean Difference is positive, both the Lower Bound and Upper
Bound are totally above zero; and wherever the Mean Difference is having
a negative sign, both the Lower Bound and Upper Bound are completely
below zero.
The participants’ Scholarly Performance Mean at the Year 1 is
significantly lower than their Scholarly Performance Mean at each of the
other four Years with [p < .0005]. The Scholarly Performance Mean at Year
2 is significantly higher than that at Year 1 and significantly smaller than
the Mean in each of the higher years [p < .0005] in each comparison. Year 3
Scholarly Performance is significantly greater than the Means at Year 1 and
Year 2, but significantly lower than the Mean at each of Years 4 and 5 [p <
.0005] in each case. Scholarly Performance Mean at Year 4 is significantly
lower than that at Year 5, but statistically greater than the Mean at each of
Years 3, 2 and 1 [p < .0005]. Finally, the Scholarly Performance Mean at
Year 5 is greater than the Mean at each of Years 4, 3, 2 and 1 with
overwhelming significance [p < .0005] in each comparison. In summary, of
the five Years, the Year 5 is the most effective with overwhelming
preponderance of Scholarly Performance Mean, followed by Year 4, and
then Year 3, and Year 2. Year 1 has the least Scholarly Performance Mean.
Multivariate Tests
The Multivariate Tests results here are exactly as those reported under the
third section of the Output 8.3. Each of the F for the four tests (1453.813) is
statistically significant with p < .0005, and has .989 Partial Eta Squared
(Effect Size). Each of the four (Pillai’s trace, Wilk’s lambda, Hotelling’s
trace and Roy’s largest root) F tests the multivariate influence of Year on
the basis of linearly independent pairwise comparisons among the estimated
marginal means.
Profile Plots
The Profile Plots is a graphic illustration of the magnitude and relative
position of each Year’s influence on the Scholarly Performance of students
in the 5-year Direct Ph.D. engineering program at the Indian Institute of
Science as shown pictorially by the Mean as well as the extent to which
583
each of the Means differs from the Observed Grand Mean (the horizontal
dark line at approximate Estimated Marginal Mean of 74). It has pictorially
visualized how effective the program was in significantly improving
scholarly Performance yearly throughout its 5-year duration. On the
average, the 5-year Direct Ph.D. engineering program effectively improved
the engineering proficiency of each admitted candidate from a Scholarly
Mean of approximately 59 in Year 1 to approximately 65 in Year 2, 75 in
Year 3, 81 in Year 4 and 91 in Year 5 as depicted pictorially by the Bar
Chart.
584
the General Linear Model Output based on Estimated Marginal Means with
Bonferroni for Adjustment of multiple comparisons was not done.
When Paired-Samples t Test is used for follow-up pairwise comparisons
of means in One-Way Within-Subjects ANOVA, the alpha level of .05 must
be adjusted downward to have the Appropriate Alpha Level Per
Comparison so that the entire set of follow-up tests does not exceed the .05
chosen alpha. The Appropriate Alpha Level Per Comparison is arrived at
by diving the overall alpha level for the set of tests (.05) by the number of
pairwise follow-up tests. In the current example with 10 pairwise
comparisons, divide .05 by 10 to have .005 (i.e. .05 ÷ 10 = .005) so that the
sum of the 10 pairwise tests does not exceed .05 alpha. That is, instead of
evaluating each pairwise comparison Sig. against .05, the p-value to use for
each test will be evaluated against an alpha level of .005. For each paired
comparison, when the p-value (Sig.) is equal to or less than .005, the null
hypothesis on the paired Years will be rejected; and when the p-value is
greater than .005, the null hypothesis of no significant mean difference
between the two Years will be retained.
Simply put, it is .005 that must be used here as the chosen alpha in
testing the tenability of each of the 10 pairwise comparisons. This is to
ensure that the probability of committing Type I error (false rejection of a
null hypothesis that should by right have been retained) does not become
greater than .05 for the entire set of the follow-up tests. The implication is
that if no adjustment were made and each of the 10 follow-up pairwise tests
are conducted at .05, the experimental Alpha Level, the overall alpha for
the set of tests would have been increased to the sum of the .05 alphas for
the 10 tests which is .500.
The Paired Samples Test of Paired Differences has shown the Mean
Difference, the Difference Std. Deviation, Std. Error Mean, and the Lower
and Upper 95% Confidence Interval of Difference, the t, df and Sig. (2-
tailed) for each of the 10 pairwise comparisons thus:
i. Year1 and Year2 Scholarly Performance (SP) has a Mean Difference
of -5.77941, t of -18.038 that is significant statistically as p < .005.
ii. Year1 and Year3 Scholarly Performance (SP) has a Mean Difference
of -16.22059, t of -44.934 that is significant statistically as p < .005.
iii. Year1 and Year4 SP has a Mean Difference of -21.69118, t of -58.587
that is significant statistically as p < .005.
585
iv. Year1 and Yaer5 SP has a Mean Difference of -31.54412, t of -75.425
that is significant statistically because p < .005.
v. Year 2 and Year3 SP has a Mean Difference of -10.44118, t of
-32.691 that is significant statistically because p < .005.
vi. Year2 and Year4 SP has a Mean Difference of -15.91176, t of -47.771
that is significant statistically as p < .005.
vii. Year2 and Year5 SP has a Mean Difference of -25.76471, t of -64.711
that is significant statistically because p < .005.
viii. Year3 and Year4 SP has a Mean Difference of -5.47059, t of -30.671
that is significant statistically as p < .005.
ix. Year3 and Year5 SP has a Mean Difference of -15.32353, t of -53.420
that is significant statistically because p < .005.
x. Year4 and Year5 SP has a Mean Difference of -9.85294, t of -46.964
that is significant statistically as the p < .005.
These 10 pairwise comparison results are exactly like the results arrived
at with the Pairwise Comparisons in the ninth section of the Output 8.3.
That is, the pairwise comparisons that have significant difference in the
Paired Samples Test of Paired Differences are exactly the ones that were
significantly different in the General Linear Model Pairwise Comparisons.
The minus signs simply show the direction that the significant Mean
Difference favours. The Mean Differences here were all arrived at by
consistently subtracting the Scholarly Performance Mean of the higher year
from that of the lower year; and since in each pairwise comparison the
mean of the higher year was greater that the Scholarly Performance mean of
the lower year, the minus sign in each case demonstrates that the significant
Mean Difference is in favour of the Scholarly Performance of the higher
year. The Scholarly Performance in Year5 was significantly better than that
in Year4, which was significantly better than that in Year3, which was itself
better significantly than that in Year2, and which was in turn significantly
better than the Scholarly Performance in Year1.
Step 6. View the entire Output by scrolling right, left, up and down the
screen. This, you have already done in the course of going through and
interpreting the output.
586
Step 7. Save the Output by clicking File and clicking Save As… or by
clicking the Save this document icon in the Toolbar to have the Save
Output As dialog box. In it, type the suitable name (Chapter 8 Output
3.spv) for the Output in the box beside File name as shown in Figure 8.44
and click Save or press Enter.
Step 8. Print the Output. Click Print icon in the Toolbar or click File
and click Print to have the Print dialog box (see Figure 8.45). Confirm that
All visible output is checked and click OK for one copy of the entire
Output 8.3 to be printed. For more copies, simply repeat the printing
operations.
587
Figure 8.45 Print dialog box
Step 9. Exit IBM SPSS Statistics by closing all SPSS windows that are open
on the screen.
Step 10. Shut down the system. Click Start, click Power, and click Shut
down.
click on (the Run selection tool in the Toolbar) when the entered
syntax is highlighted, gets the whole analysis done and results produced in
the SPSS Viewer window. Just try it out and be convinced of another and
perhaps, a better technique for execution of One-Way Repeated Measures
ANOVA.
588
Syntax 8.2 Chapter 8 Syntax 2.sps
Switch the computer on. Launch IBM SPSS Statistics by clicking its
icon on your screen, let it get ready fully and close the superimposed
menu on the Data Editor. Retrieve the data file that is called Chapter 8
Table 2 Data.sav by clicking the Open data document icon in the
Toolbar to have the Open Data dialog box (see Figure 8.46). In it, look for
and click the document name, Chapter 8 Table 2 Data.sav and press
Enter. This opens the data file as shown earlier in Figure 8.26.
589
Figure 8.46 Open Data menu
Alternatively, simply enter the data in Table 8.2 very quickly and name
as well as label the variables as was done earlier.
Click Window Go to Designated Syntax Window (see Figure
8.47); or click File New Syntax to have the Syntax Editor as
demonstrated in Figure 8.48).
590
Figure 8.48 SPSS Syntax Editor
Figure 8.49 IBM SPSS Statistics Syntax Editor with the Syntax 8.2 entered
Highlight all the entered Syntax and click the Run Selection tool
in the Toolbar to have the Output displayed in the IBM SPSS Statistics
Viewer window as presented in Output 8.4. If you are not comfortable with
591
highlighting the entire entered Syntax, no problem. Just click Run
All and SPSS will complete the analysis instantly and produce the Output
as shown in Output 8.4.
592
593
594
595
596
Pairwise Follow-up Tests
597
598
599
INTERPRETATION OF OUTPUT 8.4
Very carefully compare every unit of Output 8.4 with Output 8.3 to discern
doubtlessly that they are exactly equal. Consequently, apply the
interpretation offered for Output 8.3 to the Output 8.4.
600
Follow the remaining protocols for data analysis with the use of IBM
SPSS Statistics to eventually shut down your system.
View output
Save the output as Chapter 8 Output 4.spv and the Syntax as Chapter 8
Syntax 2.sps
601
Exit IBM SPSS
602
Chapter 9
TWO-WAY REPEATED MEASURES
ANOVA
Overview
Correct inference of causal-and-effect relationship between the independent
and dependent variables demands covariation of the events. Covariation
could either be deterministic or probabilistic. While deterministic
covariation is when manipulation of only one independent variable
produces a significant effect on the dependent variable; probabilistic
covariation occurs when two or more independent variables are
manipulated simultaneously to cause a significant effect on the dependent
variable. Very often, researchers manipulate two independent variables
simultaneously and repetitively for observation of probable deterministic
and probabilistic covariations between the independent and dependent
variables. Whenever data on one dependent variable are collected
repeatedly from a sample that is randomly drawn from a population for
determination of possible simultaneous effects of two independent
variables, the most appropriate inferential parametric statistical test for
analysing the data is Two-Way Repeated Measures Analysis of Variance.
This chapter guarantees that the user attains expert practical knowledge of
automated analysis of such data with SPSS syntax and dialog boxes point
and click methods. Equally guaranteed, is mastery of how the output can
meticulously be interpreted with most excellent ease. Effects of a new
memory drug and learning strategy on long-term memory is used for
demonstration.
603
Within-subjects effects, Estimated marginal means, Pairwise
comparisons, Two-way repeated measures ANOVA syntax,
Output interpretation, Profile plots, Memory drug, Learning
strategy, Long-term memory, Results summary.
604
Suppose that as part of the relentless efforts towards guaranteeing improved
Long-Term Memory and preventing forgetting, a company in Germany
manufactured a new memory enhancement drug, called MemDrug. The
German Psychological Society has equally been working relentlessly on the
possible use of some Learning Strategies to improve Long-Term Memory.
So, a team of three pharmacists and three psychologists was engaged by the
Government to carry out an experiment on the effects of the Memory Drug
(MemDrug) and Learning Strategy on Long-Term Memory. The team
randomly draw a sample of 20 participants for the experiment and
administered three different milligrams of MemDrug (20mg, 40mg and
60mg) as well as training on three different Learning Strategies (Dual
Coding, Retrieval Practice and Elaboration) to each of the 20 subjects for
60 days.
The administration of the MemDrug and Learning Strategy were paired
in a forward and backward order (A B C C B
A) to prevent sequencing effect and carry-over effect from
confounding the results. The pairing of the three MemDrug levels and three
Learning Strategy levels was: 20mg and Dual Coding, 20mg and Retrieval
Practice, 20mg and Elaboration; 40mg and Dual Coding, 40mg and
Retrieval Practice, 40mg and Elaboration; 60mg and Dual Coding, 60mg
and Retrieval Practice, 60mg and Elaboration. Nine measurements of Long-
Term Memory were taken, one after completion of each A B
C C B A experimental exposition session; and the Long
Term-Memory results are as tabulated in Table 9.1.
605
Research Questions
What is the effect of the new memory drug (MemDrug) on Long-Term
Memory as measured by the subjects’ means and standard deviations?
What is the effect of the Learning Strategy on Long-Term Memory as
measured by the subjects’ means and standard deviations?
What is the combined effect of Memory Drug and Learning Strategy on
Long-Term Memory as measured by the subjects’ means and standard
deviations?
606
Alternate Hypotheses
The newly manufactured memory drug (MemDrug) has a significant main
effect on Long-Term Memory.
Learning Strategy has a significant main effect on Long-Term Memory.
The MemDrug and Learning Strategy have statistically significant
interaction effect on Long-Term Memory.
Null Hypotheses
The MemDrug has no significant main effect on Long-Term Memory.
Learning Strategy has no significant main effect on Long-Term Memory.
The MemDrug and Learning Strategy have no statistically significant
interaction effect on Long-Term Memory.
Step 3. Key in the data as contained in Table 9.1 with the nine scores of
each subject constituting one row. When done, click Variable View and
type in Name and Label for the variables (treating each column of the
scores as a variable) and selecting Scale for each row under Measure
column as shown if Figure 9.1.
607
Figure 9.1 Variable View with the variables named and labelled
On completion of the Variable View, click Data View to have the entire
entered data with the actual names of the variables as presented in Figure
9.2.
Figure 9.2 Data View with the data and variables named
608
Step 4. Save the data by clicking Save this document tool in the
Toolbar to have the Save As… dialog box in which you type the File name
as illustrated in Figure 9.3, and click Save.
609
Figure 9.4 Analyze General Linear Model Repeated Measures
610
In it, type the name of the first independent variable, MemDrug, in the
Within-Subject Factor Name. Then type 3 in the Number of Levels small
box, and click Add. This action produces the screenshot in Figure 9.6.
611
Figure 9.7 Repeated Measures Define Factor(s) with MemDrug(3) and LearnStrategy(3)
added
In the box for Measure Name, type the dependent variable name,
LTMemory, shown in Figure 9.8, and click Add to have the screenshot in
Figure 9.9.
612
Figure 9.8 Repeated Measures Define Factor(s) with the Measure Name typed
613
Figure 9.9 Repeated Measures Define Factor(s) with the Measure Name added
614
Figure 9.10 Repeated Measure main dialog box
Highlight all the variables in the large box at the left by holding the Ctrl
key down and pressing the down-pointed arrow on the Keyboard. Then,
move them at once into the big box at the right just under Within-Subjects
Variables (MemDrug,LearnStrategy) by clicking the right-pointed arrow
that is directly facing the Within-Subjects Variables
(MemDrug,LearnStrategy). With this action, the dialog box will look like
the illustrated screenshot in Figure 9.11.
615
Figure 9.11 Repeated Measures main dialog box with variables moved to the right
616
Figure 9.12 Repeated Measures Profile: Plots
Highlight and move MemDrug under Factors panel at the left into
Horizontal Axis box at the right by clicking the right-pointed arrow
that is directly facing the Horizontal Axis. Next, highlight and move
LearnStrategy under the Factors field at the left into Separate Lines box
at the right by clicking the right-pointed arrow facing the Separate
Lines to have Figure 9.13.
617
Figure 9.13 Repeated Measures: Profile Plots dialog box with variables moved
accordingly
Now, you can see that near Plots has shown very clearly. Click
the that is near Plots to have the screenshot in Figure 9.14 where the
two independent variables with which the Profile Plots will be drawn have
actually been added to the analysis, and Line Chat under Chart Type has
been checked.
618
Figure 9.14 Repeated Measures: Profile Plots dialog box with variables added
619
Figure 9.15 Repeated Measures: Options
Figure 9.16 Repeated Measures Options with the two crucial operations checked
Proceed with the analysis by clicking Continue that returns you to the
Repeated Measures main dialog box. This time in the Repeated Measures
dialog box, select EM Means (Estimated Marginal Means) to have its
dialog box, visualized in Figure 9.17.
620
Figure 9.17 Repeated Measures: Estimated Marginal Means dialog box
Figure 9.18 Repeated Measures: Estimated Marginal Means with independent variables
plus their interaction transferred
621
Check the Compare main effects checkbox and Confidence
interval adjustment will show very legibly as exemplified in Figure 9.19.
Figure 9.19 Repeated Measures: Estimated Marginal Means with Compare main effects
checked
Click the small black downward arrow to have its drop-down dialog
box that contains three different tests (LSD, Bonferroni and Sidak) as
shown in Figures 9.20. Of the three tests, select Bonferroni and click it as
presented in Figure 9.21 for the Confidence interval adjustment.
622
Figure 9.20 Repeated Measures: Estimated Marginal Means small black downward arrow
for Confidence interval adjustment selected
Figure 9.21 Repeated Measures: Estimated Marginal Means with Bonferroni chosen for
the Confidence interval adjustment
623
main dialog box; and click OK for IBM SPSS Statistics to run the analysis
and produce the output in the IBM SPSS Statistics Viewer as displayed in
Output 9.1.
624
625
626
627
628
629
630
631
INTERPRETATION OF OUTPUT 9.1 – MEMORY
DRUG AND LEARNING STRATEGY
632
IBM SPSS Statistics has done elaborate analysis and provided enormous
results for the Two-Way Repeated Measures ANOVA. There are 15 sections
in the Output, 14 tables and 1 graph. Not all of these parts that are
necessarily required in the interpretation and reporting of the Two-Way
Within-Subjects ANOVA. A very brief comment is made on each of the
sections, particularly the requisite parts for answering the research
questions and mainly for testing tenability of the null hypotheses.
Within-subjects Factors
Shows a list of the 9 combinations, made up of the 3 MemDrug levels and
the 3 LearnStrategy levels of the independent variables (Factors =
MemDrug and LearnStrategy) against dependent variable (Measure =
LTMemory). Take note of the serial numbering of the pairs or combinations
and what they stand for as some of the tables will use only these pairwise
numberings. That is: 1 and 1 = Dual20mg, 1 and 2 = Retrieval20mg, and 1
and 3 = Elaboration20mg; 2 and 1 = Dual40mg, 2 and 2 = Retrieval40mg,
and 2 and 3 = Elaboration40mg; 3 and 1 = Dual60mg, 3 and 2 =
Retrieval60mg, and 3 and 3 = Elaboration60mg.
Descriptive Statistics
This table has outlined the Mean, Std. Deviation and N for each of the 9
combinations in a most easy to understand form. The Mean and Std.
Deviation respectively for Dual20mg are 43.2500 and 11.38732, for
Retrieval20mg are 39.5000 and 13.85071, for Elaboration20mg are 34.7500
and 16.01603; for Dual40mg are 47.2500 and 14.99781, for Retrieval40mg
are 45.5000 and 12.23670, for Elaboration40mg are 45.0000 and 14.86784;
and for Dual60mg are 45.0000 and 10.25978, for Retrieval60mg are
46.2500 and 11.57072, and for Elaboration60mg are 47.2500 and 10.32052.
The N for each pair is 20. The Descriptive Statistics could serve as answers
to the research questions.
Multivariate Tests
It is not mandatory for information in this table to be reported in the Two-
Way Repeated Measures results. However, reporting it cannot be said to be
outrightly out of place as it tends to indicate that both multivariate tests and
633
univariate tests could be applied for testing tenability of the null hypothesis
in question. The table has shown the Value, F, Hypothesis df, Error df, Sig.
and Partial Eta Squared arrived at on the basis of four different Multivariate
Tests (Pillai’s Trace, Wilk’s Lambda, Hotelling’s Trace and Roy’s Largest
Root) for first, MemDrug; second, LearnStrategy; and third,
MemDrug*LearnStrategy interaction in the Within-Subjects Design Exact
statistic. MemDrug has a statistically significant F of 10.802, Hypothesis df
of 2.000, Error df of 18.000, Sig. of .001, and Partial Eta Squared of .545
for each of the four Multivariate Tests. For Learning Strategy
(LearnStrategy), each of the four Multivariate Tests has an insignificant F
of 1.333, Hypothesis df of 2.000, Error df of 18.000, Sig. of .288, and
Partial Eta Squared of .129. MemDrug*LearnStrategy (the interaction effect
of MemDrug and Learning Strategy) has an insignificant F of 2.207,
Hypothesis df of 4.000, Error df of 16.000, Sig. of .114, and Partial Eta
Squared of .356. It is crystal clear that only the MemDrug that has
statistically significant effect on Long-Term-Memory (LTMemory), F =
10.802, p < .01, ɳp2 = .545, a large effect size for each of the four
Multivariate tests (Pillai’s Trace, Wilk’s Lambda, Hotelling’s Trace and
Roy’s Largest Root).
634
of freedom for the average tests of significance in the Test of Within-
Subjects Effects.
The Mauchly’s Test of Sphericity has shown that for MemDrug main
effect, Sphericity Assumed should be used because Mauchly’s W (.918) is
not significant (Sig. = .463) as this p > .05. For LearnStrategy main effect,
Sphericity Assumed should be used because Mauchly’s W (.866) is not
significant (Sig. = .275) because this p > .05. Similarly, for
MemDrug*LearnStrategy interaction effect, Sphericity Assumed should be
used because Mauchly’s W (.541) is not significant (Sig. = .299) because
this p > .05. However, as I have commented earlier, when Mauchly’s Test of
Sphericity was interpreted in the previous chapter, it is safer to also provide
results based on Greenhouse-Geisser as an additional alternative test in
reporting the Two-Way Repeated Measures ANOVA Tests of Within-
Subjects Effects.
635
Long-Term-Memory of individuals in the population from which the
representative sample was drawn for the study.
Using Greenhouse-Geisser (if the Sphericity Assumption had been
violated), MemDrug has Type III Sum of Squares of 1892.500, df of 1.849,
Mean Square of 1023.777, F of 13.173, Sig. of .000 (i.e., less than .0005),
and Partial Eta Squared (Effect Size) of .409. The Error(MemDrug) has
Type III Sum of Squares of 2729.722, df of 35.122, and 77.720 Mean
Square. Since the Sig. or p-value is less than the chosen alpha (be it at .05
or .01), the null hypothesis that “the new Memory Drug (MemDrug) has no
significant main effect on Long-Term Memory” is rejected [F(1.849,
35.122) = 13.173, p < .01, ɳp2 = .409 (a large effect size)]. The rejection of
this null hypothesis assuredly shows that Long-Term Memory significantly
differs across the three MemDrug dosage conditions (20mg, 40mg, and
60mg). Another test known as Pairwise Comparisons will clarify this main
effect shortly.
It should be recalled from the previous three chapters that Partial Eta
Squared (ɳp2) is the measure of Effect Size in ANOVA models. Also, that
classification of the Partial Eta Squared as the measure of Effect Size in
Analysis of Variance (ANOVA) models has been given as follows:
.000 - .009 = Trivial Effect Size
.010 - .059 = Small Effect Size
.060 - .139 = Medium Effect Size
.140 and above = Large Effect Size.
Using Sphericity Assumed, LearnStrategy has 240.833 Type III Sum of
Squares, 2 df, 120.417 Mean Square, 1.921 F, .160 Sig., and .092 Partial
Eta Squared. Its Error(LearnStrategy) has 2381.389 Type III Sum of
Squares, 38 df, and 62.668 Mean Square. Since the Sig. (.160) is greater
than the chosen alpha of .05, the null hypothesis that “Learning Strategy has
no significant main effect on Long-Term Memory” is retained [F(2, 38) =
1.921, p > .05, ɳp2 = .092, a medium effect size]. Indeed, LearnStrategy
does not have significant main effect on Long-Term Memory. Though the
effect of Learning Strategy (LearnStrategy) on Long-Term-Memory is not
statistically significant, 9.2% of the variance in Long-Term-Memory is
accounted for by the Learning Strategy (Dual Coding, Retrieval Practice
and Elaboration) adopted by the individuals.
The same conclusion is made when Greenhouse-Geisser is used as
LearnStrategy has 240.833 Type III Sum of Squares, 1.764 df, 136.516
636
Mean Square, 1.921 F, .166 Sig. and .092 Partial Eta Squared. Also, with
Greenhouse-Geisser, LearnStrategy has Error(LearnStrategy) 2381.389
Type III Sum of Squares, 33.519 df, and 71.047 Mean Square. The
computed Sig. (.166) is greater than the chosen alpha of .05. Therefore,
retain the null hypothesis that “Learning Strategy has no significant main
effect on Long-Term Memory” [F(1.764, 33.519) = 1.921, p > .05, ɳp2 =
.092 (medium effect size)].
With Sphericity Assumed, MemDrug*LearnStrategy has 591.667 Type
III sum of Squares, 4 df, 147.917 Mean Square, 2.442 F, .054 Sig., and .114
Partial Eta Squared. Its Error(MemDrug*LearnStrategy) has 4602.778 Type
III Sum of Squares, 76 df, and 60.563 Mean Square. The Sig. (.054) is
greater than the chosen alpha of .05, therefore the null hypothesis that
“Memory Drug and Learning Strategy have no statistically significant
interaction effect on Long-Term Memory” is retained (not rejected) [F(4,
76) = 2.442, p > .05, ɳp2 = .114 (medium effect size)].
If the results are viewed on the basis of Greenhouse-Geisser (as though
the Sphericity Assumption were not met), MemDrug*LearnStrategy has
591.667 Type III Sum of Squares, 2.912 df, 203.185 Mean Square, 2.442 F,
.075 Sig., and .114 Partial Eta Squared. The
Error(MemDrug*LearnStrategy) has 4602.778 Type III Sum of Squares,
55.327 df, and 83.192 Mean Square. Since the Sig. (.075) is greater than the
chosen alpha of .05, the null hypothesis that there is no significant
interaction effect of the new Memory Drug (MemDrug) and Learning
Strategy on Long-Term Memory is retained (not rejected) [F(2.912, 55.327)
= 2.442, p > .05, ɳp2 = .114, a medium effect size]. Although the
interaction effect of the new Memory Drug (MemDrug) and Learning
Strategy (LearnStrategy) on Long-Term Memory is not significant
statistically, 11.4% of the variance in Long-Term Memory is accounted for
by the MemDrug and LearnStrategy interaction. If 11.4% of the variance in
Long-Term Memory is considered as a meaningful gain, then it would not
be out of place for both the Memory Drug and Learning Strategy to be
recommended for use.
637
outcome of two trend tests known as Linear and Quadratic with respect to
the effects of MemDrug and Learning Strategy, as well as of the
MemDrug*LearnStrategy interaction, on Long-Term Memory
(LTMemory). For instance, MemDrug has a significant Linear trend effect
on Long-Term Memory: [F(1, 19) = 15.990, p < .01, ɳp2 = .457]. That is,
the points represented by the effect of MemDrug on Long-Term Memory as
depicted by the means, actually significantly tend to fall onto a straight line.
MemDrug also has significant Quadratic trend effect on Long-Term
Memory: [F(1, 19) = 8.166, p < .05, ɳp2 = .301]. That is, the points
represented by the effect of MemDrug on Long-Term Memory as depicted
by the means, significantly tend to fall under an inverted U-shaped trend.
The Linear and Quadratic trend tests show reasonable consistency in the
trend of MemDrug effect on Long-Term Memory. They indicate that over
time, MemDrug has meaningfully predictable improvement effect on Long-
Term Memory.
In all, while a Linear trend test is used to assess the extent to which the
points, represented by the means, tend to fall onto a straight line; a
Quadratic trend test assesses the extent to which the points, represented
by the means, tend to fall under a U-shaped or an inverted U-shaped trend
as can be seen pictorially in the Profile Plots section of the output.
Furthermore, the Tests of Within-Subjects Contrasts as in this table mainly
indicates whether a trend is significant or not significant. It has shown
whether the individual conditions significantly differ from one another in a
meaningful pattern or in a haphazard form. When the trend test is
significant as found with the effect of MemDrug on Long-Term Memory,
the trend or pattern of the effect is meaningful, and not by mere coincidence
or chance. Another test, known as Pairwise Comparisons, to be discussed a
short while from now, will further explicitly indicate the extent to which
individual Long-Term Memory conditions significantly differ from one
another as a function of the MemDrug.
LearnStrategy does not have a significant Linear trend effect on Long-
Term Memory: [F(1, 19) = 2.814, p > .05, ɳp2 = .129]. That is, the points
represented by the effect of LearnStrategy on Long-Term Memory as
depicted by the means, do not significantly tend to fall onto a straight line
as can be seen graphically in the Profile Plots. Also, LearnStrategy does
not have a significant Quadratic effect on Long-Term Memory: [F(1, 19) =
638
.000, p > .05, ɳp2 = .000]. That is, the points represented by the effect of
LearnStrategy on Long-Term Memory as depicted by the means, do not
significantly tend to fall under a U-shaped or an inverted U-shaped line.
The lack of significance in the Linear and Quadratic trends of the effect of
LearnStrategy on Long-Term Memory depicts haphazard or chance
occurrence that cannot be meaningfully predicted.
For Linear, while the MemDrug*LearnStrategy interaction Linear trend
effect on Long-Term Memory is significant: [F(1, 19) = 9.029, p < .01, ɳp2
= .322]; MemDrug*LearnStrategy interaction Quadratic trend effect on
Long-Term Memory is insignificant: [F(1, 19) = .023, p > .05, ɳp2 = .001].
For Quadratic, while the MemDrug*LearnStrategy interaction Linear trend
effect on Long-Term Memory is insignificant: [F(1, 19) = .074, p > .05, ɳp2
= .004]; MemDrug*LearnStrategy interaction Quadratic trend effect on
Long-Term Memory is also insignificant: [F(1, 19) = .115, p > .05, ɳp2 =
.006].
639
MemDrug information could be used partly to answer the first research
question.
640
Difference. Where the mean of the (I) MemDrug is greater than the mean of
the (J) MemDrug, there is no negative sign as prefix of the Mean
Difference.
The evidence of statistical significance in each Mean Difference can also
be seen in the 95% Confidence Interval for Difference Lower Bound and
Upper Bound because wherever the significant Mean Difference is positive,
both the Lower Bound and Upper Bound are totally above zero; and
wherever the significant Mean Difference is having a negative sign, both
the Lower Bound and Upper Bound are completely below zero. If the Mean
Difference is not statistically significant, the Lower Bound and the Upper
Bound of 95% Confidence Interval for Difference must have zero between
them.
Multivariate Tests
These Multivariate Tests results are exactly as those reported under the third
section of the Output 9.1 for MemDrug. Each of the F for the four tests
(10.802) is statistically significant with p < .05 and p < .01, and has .545
Partial Eta Squared (Effect Size). Each of the four tests (Pillai’s trace,
Wilk’s lambda, Hotelling’s trace and Roy’s largest root) uses Exact
statistic F to test the multivariate effect of MemDrug on the basis of
linearly independent pairwise comparisons among the estimated marginal
means.
641
LearnStrategy Estimated Marginal Means could be used for partly
answering the second research question.
642
Long-Term Memory Mean Difference is 2.833, Std. Error is 1.689, Sig. is
.329 (p > .05), the Lower Bound and Upper Bound at 95% Confidence
Interval for Difference are -1.600 and 7.267, respectively. Similar
interpretations go for the other pairwise comparisons in the table.
To show the direction that the non-significant Mean Difference favours,
where the LTMemory mean of the (I) LearnStrategy is smaller than the
LTMemory mean of the (J) LearnStrategy, a negative sign (-) is put as the
prefix of the given Mean Difference because the (J) mean is subtracted from
the (I) mean to arrive at the Mean Difference. Where the LTMemory mean
of the (I) LearnStrategy is greater than that of the mean of the (J)
LearnStrategy, there is no negative sign as prefix of the Mean Difference.
The evidence of absence of statistical significance in each Mean
Differences can also be seen in the 95% Confidence Interval for Difference
Lower Bound and Upper Bound because for every of the Mean Difference,
both the Lower Bound and Upper Bound spread below zero and above zero.
Multivariate Tests
Results of the Multivariate Tests here are exactly as those reported under
the third section of the Output 9.1 for LearnStrategy. Each of the F for the
four tests (1.333) is not significant as the Sig. (.288) is greater than the
chosen .05 level of significance [p > .05], and has only .129 Partial Eta
Squared (Effect Size). Each of the four tests (Pillai’s trace, Wilk’s lambda,
Hotelling’s trace and Roy’s largest root) uses Exact statistic F to test the
multivariate effect of Learning Strategy on the basis of linearly independent
pairwise comparisons among the estimated marginal means.
643
Retrieval60mg, and 3 and 3 = Elaboration60mg). The Dual20mg has a
Mean of 43.250 and Std. Error of 2.546; 95% Confidence Interval Lower
Bound of 37.921 and Upper Bound of 48.579. Similarly, the Retrieval20mg
has a Mean of 39.500 and Std. Error of 3.097; 95% Confidence Interval
Lower Bound of 33.018 and Upper Bound of 45.982. The Elaboration20mg
has a Mean of 34.750 and Std. Error of 3.581; 95% Confidence Interval
Lower Bound of 27.254 and Upper Bound of 42.246. Equally, the
Dual40mg has a Mean of 47.250 and Std. Error of 3.354; 95% Confidence
Interval Lower Bound of 40.231 and Upper Bound of 54.269. Similar
interpretation goes for each of the remaining pairs of the Memory Drug and
Learning Strategy interaction with regards to Long-Term Memory Means,
Std. Error and 95% Lower Bound and Upper Bound Confidence Interval.
The MemDrug*LearnStrategy interaction information in this section of the
Output 9.1 could partly be used to answer the third research question.
Profile Plots
The Profile Plots has graphically illustrated the magnitude and relative
position of the nine means (three for MemDrug, three for LearnStrategy,
and three for the interaction between MemDrug and LearnStrategy
[MemDrug*LearnStrategy]) with regards to Long-Term Memory. The line
for Dual Coding LearnStrategy runs through MenDrug20mg at Long-Term
Memory mean of 43.25, via MenDrug40mg at LTMemory mean of 47.25 to
MemDrug60mg at LTMemory mean of 45.00. The line for Retrieval
Practice LearnStrategy runs through MenDrug20mg at Long-Term Memory
mean of 39.50, via MenDrug40mg at LTMemory mean of 45.50 to
MemDrug60mg at LTMemory mean of 46.25. The line for Elaboration
LearnStrategy runs through MemDrug20mg at Long-Term Memory mean
of 34.75, via MemDrug40mg at LTMemory mean of 45.00 to
MemDrug60mg at LTMemory mean of 47.25.
644
pharmaceutical company and Learning Strategies on Long-Term Memory
showed a significant main effect of the MemDrug [F(2, 38) = 13.173, p <
.01, ɳp2 = .409, a large effect size] based on Sphericity Assumed. The
Memory Drug equally has a significant main effect on Long-Term Memory
[F(1.849, 35.122) = 13.173, p < .01, ɳp2 = .409, a large effect size] based
on Greenhouse-Geisser. The significant difference is in favour of
MemDrug60mg (EM Mean = 46.167) p <.01, and MemDrug40mg (EM
Mean = 45.917) p < .01. The main effect of Learning Strategy on Long-
Term Memory is not significant [F(2, 38) = 1.921, p > .05, ɳp2 = .092,
medium effect size] based on Sphericity Assumed. On the basis of
Greenhouse-Geisser too, the main effect of Learning Strategy is not
significant [F(1.764, 33.519) = 1.921, p > .05, ɳp2 = .092, medium effect
size]. The interaction effect of Memory Drug and Learning Strategy on
Long-Term Memory is not significant [F(4, 76) = 2.442, p > .05, ɳp2 =
.114, a medium effect size] based on Sphericity Assumed. Using
Greenhouse-Geisser too, the interaction effect (MemDrug*LearnStrategy)
is not significant [F(2.912, 55.327) = 2.442, p > .05, ɳp2 = .114, a medium
effect size]. Therefore, for improvement of Long-Term Memory, MemDrug
40mg is recommended. For better improvement of Long-Term Memory, the
MemDrug 60mg is recommended.
Step 6. View the output, which has already been done in the interpretation
process.
Step 7. Save the output as Chapter 9 Output 1.spv. Click Save this
document tool in the Toolbar to have the Save Output As… dialog
box. In the box beside File name, type in a suitable name for the document
(Chapter 9 Output 1.spv) as shown in Figure 9.22; and click Save.
645
Figure 9.22 Save Output As… Chapter 9 Output 1.spv
Step 8. Print the output and the dataset. Click the Print icon in the
Toolbar to get the Print dialog box. Ensure that All visible output is
selected as shown in Figure 9.23 and click OK to have one copy of the
entire output for the Two-Way Repeated Measures ANOVA printed. After
that, activate the SPSS Data Editor with the dataset as shown in Figure 9.2
and click the Print icon and click Ok to have a copy of the data printed.
646
Figure 9.23 Print dialog box
Step 9. Exit IBM SPSS Statistics by closing each opened SPSS window.
Step 10. Shut down the system by clicking Start, clicking Power and
finally by clicking Shut down.
entered syntax and clicking the Run Selection tool in the Toolbar.
With that click, IBM SPSS Statistics completes the analysis and produces
the output.
647
Without further ado, switch on the computer and let it boot fully. Launch
IBM SPSS Statistics by clicking its icon on the screen of the
computer. When it fully comes up, close the Welcome to IBM SPSS
Statistics dialog box. Move cursor to the Toolbar and click the Open data
document tool to have its menu as given in Figure 9.24. In the Open
Data menu, look for and click the name with which the data document was
saved (Chapter 9 Table 1 Data.sav) and press Enter key on the Keyboard
or click Open.
648
Figure 9.24 Open Data for Chapter 9 Table 1 Data.sav
With the pressing of the Enter key, the document opens exactly as
shown much earlier in Figure 9.2. Have a very close look at the variables as
named because the names are indispensable part of the Syntax.
Select Window and click Go to Designated Syntax Window (see
Figure 9.25) or click File, select New and click Syntax.
649
Figure 9.26 IBM SPSS Statistics Syntax Editor
Most carefully enter the Syntax that is presented in Syntax 9.1. When
done, it will look like the screenshot in Figure 9.27.
Figure 9.27 IBM SPSS Statistics Syntax Editor with the requisite Two-Way Repeated
Measures ANOVA syntax
Crosscheck to verify that you have correctly typed the Syntax as given
in Syntax 9.1. With exact names of the variables as you have carefully
noted in Figure 9.2, the clear specification of the Syntax as in Syntax 9.1,
and with the aid of the prompting that IBM SPSS Statistics gives while
typing the Syntax, probability of your making a mistake is very remote. Do
not miss out the dot (.) or full stop at the end of the command Syntax. The
command syntax for each statistical analysis must end with a dot.
Finally, highlight the entire entered Syntax and click the Run Selection
icon in the Toolbar for SPSS to execute the whole analysis and
produce the results in a second as in the displayed Output 9.2.
650
Alternatively, instead of highlighting the entered Syntax, click Run in
the Menu Bar and in the Run dropdown menu, click All (see Figure 9.28).
With this, IBM SPSS Statistics instantly runs the analysis and presents the
results in the Viewer window as exhibited in Output 9.2.
651
652
653
654
655
656
657
658
659
INTERPRETATION OF OUTPUT 9.2
Compare every table in Output 9.2 with its corresponding table in Output
9.1 and make a list of the differences. At the end, how many differences
660
were you able to spot and list? Zero difference? Yes. Good! If there is no
difference at all in the two outputs (Output 9.2 and Output 9.1), then the
explanations provided for Output 9.1 cannot be different from how Output
9.2 should be explained. Apply the interpretation of Output 9.1 to the
Output 9.2.
Perform the remaining IBM SPSS Statistics data analysis protocols to
eventually shut down your system.
View output
Save the output (Chapter 9 Output 2.spv) and the Syntax (Chapter 9
Syntax 1.sps)
661
Chapter 10
TWO-WAY MIXED (BETWEEN-
WITHIN) ANALYSIS OF VARIANCE
Overview
In this chapter, 44 screenshots demonstration of unmistakable step-by-step
hands-on knowhow for the execution of Two-Way (Between-Within)
Analysis of Variance is presented. This statistical test is the best for
concurrent ascertainment of the effects of one or two between-subjects
factors and one or two within-subjects factors on a single dependent
variable. While the between-subjects factor is a categorical independent
variable, the within-subjects factor is a multi-treatment independent
variable that considers all the subjects as a single sample. Two-way
(between-within) ANOVA is a much more powerful factorial design test
with greater precision in the specification of within-subjects main effect,
between-subjects main effect, within-subjects and between-subjects
interaction effect on one dependent variable. Multivariate Analysis of
Variance (MANOVA) for follow-up of the interaction effect is equally
presented in an extremely simplified user-friendly manner. The illustrative
example used is staff recruitment exercise for a STEM program in gifted
children’s school.
662
at least one between-subjects factor and at least one within-subjects factor
as the independent variables on a dependent variable. The Between-
Subjects factor is an independent variable with preferably three or more
distinct levels or groups in which members of each group were accorded
only one level of the treatment condition. The Within-Subjects factor is an
independent variable with preferably three or more treatment conditions in
which each of the subjects was exposed to all the treatment conditions. That
is, the Two-Way Mixed (Between-Within) ANOVA is as the name implies,
a mixed factorial ANOVA design that combines between-subjects and
within-subjects experimental conditions simultaneously in a single study
and analysis. An example of how the analysis is done will clarify the mixed
factorial ANOVA design.
663
mathematics (arithmetic, applied maths, calculus, algebra, experimental
maths, logic, geometry/topology, dynamic systems with differential
equations, and data mining). Science training dealt with areas that include
biochemistry, biology, medicine, and psychology. Technology training
touched a broad spectrum of the field, including artificial intelligence,
curriculum development and educational technology, information and
communication technology, computer science and programming. The
intention was to at the end of the training, recruit only those candidates who
will excel in a pre-employment STEM Test Battery (STEMTB) that will be
given, to serve as tutors in the STEM-based Gifted Children’s School. The
STEMTB was adopted for the measurement of the Staff Recruitment
Exercise Performance (SREP).
The four core areas of training, Science, Technology, Engineering, and
Mathematics (STEM) constituted the Within-Subjects Factor in which
every of the candidates was equally exposed to all the four levels. The areas
of Discipline in which each candidate got his/her first degree served as the
Between-Subjects Factor because each of the candidates had a degree in
just one Discipline.
The STEM Test Battery (STEMTB) was a pre-employment set of four
tests grouped together and administered to the participants in the training
for possible employment as teaching staff in the STEM-based Gifted
Children’s School that was to be established by the Governor. In accordance
with the requirements of the job and the employer’s needs, the four tests in
STEMTB were Science, Technology, Engineering and Mathematics for the
measurement of Staff Recruitment Exercise Performance (SREP). Each of
these tests had in addition to its core area, some items that measured basic
skills and trainability as well as personality predispositions of creativity,
trustworthiness, honesty, and adherence to rules.
The four tests in the STEMTB were administered in a forward and
backward sequence (A B C D D C
B A) and the average of each applicant’s two scores on each of the
tests was taken as his/her score for the particular test. The A B
C D D C B A order of administration of
the Within-Subjects Factor was to prevent carry-over effect and sequencing
effect (Ololube, Kpolovie, & Makewa, 2015; Kpolovie, 2016; 2010) from
664
affecting the findings. Once again, the Staff Recruitment Exercise
Performance was measured with the STEMTB. Three Disciplines that had
the highest number of applicants were used for the study, and coded 1 for
MBBS, 2 for Agriculture, and 3 for Management Science. The
applicants’ scores are as presented in Table 10.1.
665
666
Research questions
What is the applicants’ staff recruitment exercise performance (SREP) in
terms of their areas of discipline as depicted by their means and standard
deviations?
667
What is the applicants’ staff recruitment exercise performance (SREP) in
terms of the STEM training that they received as depicted by their means
and standard deviations?
What is the applicants’ staff recruitment exercise performance (SREP)
with regards to the interaction of their discipline and STEM training?
Alternate Hypotheses
The applicants’ performance in the staff recruitment exercise differ
significantly with regards to their discipline.
The applicants’ performance in the staff recruitment exercise differ
significantly with regards to the STEM training that they received.
There is a significant interaction effect of discipline and STEM training
on the applicants’ performance in the staff recruitment exercise.
Null hypotheses
The applicants’ performance in the staff recruitment exercise do not differ
significantly with regards to their discipline.
The applicants’ performance in the staff recruitment exercise do not
differ significantly with regards to the STEM training that they received.
There is no significant interaction effect of discipline and STEM training
on the applicants’ staff recruitment exercise performance.
Step 2. Launch IBM SPSS Statistics by clicking the icon of the version of
SPSS that is installed in your computer. Version 26 for instance is like this
. When fully ready, close the superimposed dialog
box on the IBM SPSS Data Editor.
Step 3. Enter the data exactly as shown in Table 10.1 with the five scores
for each applicant constituting one row in the same order. When the entire
668
scores for the 60 applicants have correctly been entered exactly the way
they appear in Table 10.1, switch to the Variable View by clicking it beside
Data View at the bottom extreme left corner of the SPSS Statistics Data
Editor. In the Variable View, under the Name column, type the actual
names of the variables (Science, Technology, Engineering, Maths, and
Discipline) respectively to replace the default names that SPSS has given
them (VAR00001, VAR00002, VAR00003, VAR00004, and VAR00005).
Also enter the variable names (Science, Technology, Engineering, Maths,
and Discipline) accordingly in the first five rows under Label column. In
the Measure column, select Scale for the first four rows, and Nominal for
the fifth row because Discipline, the Between-Subjects independent
variable is categorized into three (coded 1, 2 and 3).
Then for the fifth row under Values, click None, and click the small blue
dotted box . This produces its dialog box known as Value Labels as
presented in Figure 10.1.
Type 1 in the box beside Value and type MBBS in the box beside Label
as shown in Figure 10.2.
669
Figure 10.2 Value Label with 1 and MBBS entered
Click Add, after which you type 2 in the Value box; and type Agric in
the Label box as illustrated in Figure 10.3.
Click Add, after which you type 3 in the Value box; and enter MS
(standing for Management Sciences) in the Label field as shown in Figure
10.4.
670
Figure 10.4 Value Labels with 3 and MS entered
Finally, click Add to complete the Value Labels dialog box for it to look
like the screenshot illustration in Figure 10.5.
Then, click OK, to return to the Variable View when all variables have
been named, labelled and the Between Factor independent variable
(Discipline) has been given Value Labels as shown in Figure 10.6.
671
Figure 10.6 Completed Variable View with all variable names and labelled
Switch over to the Data View again to see the entered data and actual
variable names instead of the default of VAR00001, VAR00002, and so on,
that was given when you keyed in the data. The completed Data View of
the IBM SPSS Data Editor with the data is presented in Figure 10.7.
672
Figure 10.7 Chapter 10 Table 1 Data.sav
673
Figure 10.7 Continued
674
Figure 10.7 Continued
Step 4. Save the data. Move cursor to the Toolbar and click Save this
document icon . With this click, the Save Data As dialog box appears.
In it, type a most suitable name, Chapter 10 Table 1 Data.sav, in the File
name box as exemplified in Figure 10.8, and click Save.
675
Figure 10.8 Save Data As… Chapter 10 Tata 1 Data.sav
Step 5. Analyze the data. This is the step in which all the required statistical
operations for execution of the Two-Way Mixed (Between-Within)
ANOVA have to be done. Click Analyze General Linear Model
Repeated Measures… as illustrated in Figure 10.9.
676
On clicking the Repeated Measures…, a dialog box that is called
Repeated Measures Define Factor(s) appears, shown in Figure 10.10. For
the Within-Subject Factor Name, STEM, (referring to the intensive
STEM training that was accorded to all the applicants as the experimental
treatment conditions); and the Number of Levels in this treatment Factor,
4 (i.e., 1. Science, 2. Technology, 3. Engineering and 4. Mathematics) to be
typed in their corresponding boxes and added to the analysis by clicking
Add. Also, to be typed in is the dependent variable, termed here as
Measure Name, SREP (synonym for Staff Recruitment Exercise
Performance) and added by clicking Add. Perform these actions, one after
the other.
677
uppercase (capital letter) as the beginning alphabet of the Factor Name in
some versions of SPSS. Therefore, Stem is typed instead of STEM.
Figure 10.11 Repeated Measures Define Factor(s) with factor1 name entered
678
Figure 10.12 Repeated Measures Define Factor(s) with the number of levels in the Within-
Subjects factor provided
Next, click Add, for the name of the Within-Subject Factor(s), Stem,
and its number of levels, 4, to be added to the analysis as shown in Figure
10.13.
679
Figure 10.13 Repeated Measures Define Factor(s) with the factor and its levels added
680
Figure 10.14 Repeated Measures Define Factor(s) with Measure Name entered
681
Figure 10.15 Repeated Measures Define Factor(s) with the dependent variable Srep (Staff
Recruitment Exercise Performance) added
682
Figure 10.16 Repeated Measures dialog box
683
box. When all the actions to be taken for the analysis are done, it is still in
the Repeated Measures dialog box that the final button, OK, will be
clicked for the IBM SPSS Statistics to run the analysis and produce the
much-needed output. Alright, perform these necessary statistical operations
one after the other carefully.
Hold down the Shift key and press the Arrow-down key on the
Keyboard to select all the levels of the Within-Subjects Factor or of the
independent variable that are each treated as a variable (Science,
Technology, Engineering and Maths) in the left box and move them at
once into the Within-Subjects Variables box at the right by clicking the
right-pointed arrow that is facing it as shown in Figure 10.17.
Figure 10.17 Repeated Measures dialog box with the Within-Subjects variables moved to
the right
684
the right by clicking the right-pointed arrow that is directly facing the
Between-subjects Factor(s) box as exhibited in Figure 10.18.
Figure 10.18 Repeated Measures dialog box with Discipline moved to Between-Subjects
Factor(s) box
685
Figure 10.19 Repeated Measures: Profile Plots dialog box
Click Stem in the Factors box at the left to highlight it. Then, transfer it
into the Horizontal Axis by clicking the right-pointed arrow that is
directly facing the Horizontal Axis. It is a rule of thumb for the variable
with greater number of levels (categories) to be transferred into the
Horizontal Axis and the variable with fewer number of levels to be
transferred into the Separate Lines (Kpolovie, 2018). The within-subjects
independent variable, Stem, has four levels (Science, Technology,
Engineering and Mathematics) while Discipline has three levels (MBBS,
Agriculture and Management Sciences). This accounts for why the Stem
686
is moved into the Horizontal Axis. Next, click Discipline in the Factors
box at the left to highlight it; and move it into the Separate Lines by
clicking the right-pointed arrow that is facing the Separate Lines
directly as demonstrated in Figure 10.20.
Figure 10.20 Repeated Measures: Profile Plots with variables transferred to two boxes at
the right
687
. A click on the Add produces the interaction term Stem*Discipline
that is indispensably required for either Line chart or Bar chart as
illustrated in Figure 10.21. Also, you must notice that by clicking the Add,
the two independent variables, Stem and Discipline, disappeared from their
respective Horizontal Axis and Separate Lines fields, and rather appeared
as an interaction term, Stem*Discipline in the big panel under Plots.
688
Figure 10.22. The Post Hoc statistical operation in Two-Way Mixed
(Between-Within) ANOVA is used for Post Hoc Multiple Comparisons
of the dependent samples means strictly with respect to the Between-
Subjects independent variable, which is Discipline in this example.
Move Discipline from the Factor(s) box at the left into the Post Hoc
Tests for panel at the right by clicking the right-pointed arrow, , when
the variable in question is highlighted. The rightward movement of the
variable activates the several Post Hoc Multiple Comparisons tests as
exhibited in Figure 10.23.
689
Figure 10.23 Repeated Measures: Post Hoc when Discipline has been moved to the right
Check the checkbox of the particular Post Hoc test(s) suitable for the
purpose that you want to use. Check R-E-G-W-Q (that is very suitable
when equal variances is assumed) and Games-Howell (that is appropriate
in case equal variances is not assumed) as demonstrated in Figure 10.24.
690
Figure 10.24 Repeated Measures: Post Hoc tests checked
691
Figure 10.25 Repeated Measures: Estimated Marginal Means dialog box
With the aid of Shift key and Arrow-down key on the Keyboard, select
Discipline, Stem and Discipline*Stem in the Factor(s) and Factor
Interactions box at the left and transfer them into the Display Means for
box at the right by clicking the right-pointed arrow . You can also
highlight each of the three variables in the left box and transfer it separately
into the right panel by clicking the right-pointed arrow. When done, the
Repeated Measures: Estimated Marginal Means dialog box takes a
different look as in Figure 10.26.
692
Figure 10.26 Repeated Measures: Estimated Marginal Means dialog box with the main
effects and interaction effect moved to the right
Figure 10.27 Repeated Measures: Estimated Marginal Means dialog box with Compare
main effects ticked and Confidence interval adjustment activated
693
Click the small downward-pointed black arrow under Confidence
interval adjustment and click Bonferroni in its dropdown menu for the
Confidence interval adjustment as illustrated in Figure 10.28.
Figure 10.28 Estimated Marginal Means with Bonferroni chosen for Confidence interval
adjustment
Click Continue to get back to the Repeated Measures main dialog box
for execution of the last set of statistical operations. In the main Repeated
Measures dialog box, select Options pushbutton. This action produces the
Repeated Means: Options presented in Figure 10.29.
694
This dialog box allows you to perform descriptive statistics and the so
much needed Effect Size, measured with Partial Eta Squared, when any
brand of Two-way ANOVA is performed. You get the descriptive statistical
values like means and standard deviations, and the Effect Size analysis
done merely by checking the checkbox beside Descriptive statistics and
the checkbox beside Estimates of effect size as presented in Figure 10.30.
Figure 10.30 Repeated Measures: Options with Estimates of effect size and Descriptive
statistics checked
695
Figure 10.31 Repeated Measures main dialog box when all the statistical operations have
been done
Move the cursor to, and click OK, and in just a moment, IBM SPSS
Statistics runs the whole analysis completely and displays the output in the
IBM SPSS Statistics Viewer as presented in Output 10.1.
696
with the three levels of Discipline (the Between-Subjects factor) has
significant effect on the dependent variable (Srep). If the interaction effect
of each Stem level with all the three levels of Discipline is significant on
Srep, at least one of the means of the interaction must be statistically
different from one or some of the others. In synopsis, the MANOVA will
reveal whether a significant mean difference exists across the Between-
Subjects factor (Discipline) for each Within-Subjects factor (Stem) level,
on the one hand.
From the other perspective, the MANOVA will show whether the
interaction between each of the three levels of the Between-Subjects factor
(Discipline) with all the four levels of the Within-Subjects factor (Stem) is
significant statistically. If at least one of the interactions of each Discipline
level with across all levels of Stem is greater statistically than one or more
of the others, the MANOVA will yield a significant result. In synopsis, the
MANOVA will reveal whether a significant mean difference exists across
the Within-Subjects factor (Stem) levels for each level of the Between-
Subjects factor (Discipline) level, and viseveser.
Execution of MANOVA to test the null hypothesis of ‘no significant
mean difference across the Between-Subjects factor (Discipline) levels for
each Within-Subjects factor (Stem) level”; and the execution of
MANOVA to test the null hypothesis of “no significant mean difference
across the Within-Subjects factor (Stem) levels for each Between-
Subjects factor (Discipline) level” are done with the two respective IBM
SPSS Statistics Syntax Commands given in Syntax 10.1.
697
Note that there are two Syntax Commands, each begins with MANOVA
and ends with a dot (full stop). The first Syntax is used to test the null
hypothesis that “there is no significant mean difference across the Between-
Subjects factor (Discipline) levels for each Within-Subjects factor
(Stem) level.” The second Syntax is used for testing tenability of the null
hypothesis that “there is no significant mean difference across the Within-
Subjects factor (Stem) levels for each Between-Subjects factor
(Discipline) level.”
Activate the IBM SPSS Statistics Data Editor with the dataset that was
saved as Chapter 10 Table 1 Data.sav. Click Window, and click Go to
Designated Syntax Window. Alternatively, click File, select New, and
click Syntax. Either of the two processes will instantly make the IBM
SPSS Statistics Syntax Editor to appear.
Correctly enter the Syntax that is provided in Syntax 10.1 for simple
Between-Subjects * Within-Subjects interaction effect into the IBM
SPSS Statistics Syntax Editor as exemplified in Figure 10.32.
698
Figure 10.32 Chapter 10 Syntax 1.sps – MANOVA Syntax for further Stem * Discipline
Interaction follow-up tests
Click Run. Click All. With this click on All, IBM SPSS Statistics
immediately runs the MANOVA internally and produces the results in the
IBM SPSS Statistics Viewer as shown in the follow-up section of Output
10.1.
699
700
701
702
703
704
705
706
RESULTS OF MANOVA FOR INTERACTION
FOLLOW-UP
707
708
709
710
INTERPRETATION OF OUTPUT 10.1 – EFFECTS
OF STEM AND DISCIPLINE
The Output 10.1 is very extensive with 18 sections that are made up of 17
tables and 1 graph; though, not all the 18 sections that are necessarily
required for reporting the findings. In addition, there are 10 sections of the
Output as MANOVA follow-up for the significant interaction effect
between Discipline and STEM. Each of the sections in the Output is aptly
detailed herein with the essential interpretation.
Within-Subjects Factors
This merely lists the four within-subjects treatment conditions and
measurements or observations (Science, Technology, Engineering and
Mathematics) taken across all the subjects.
Between-Subjects Factors
The Value Labels for the three levels (MBBS, Agric and Management
Science) of the Between-Subjects variable, Discipline, and the number of
cases in each level are outlined in this table.
Descriptive Statistics
The Mean, Std. Deviation and sample size (N) for each of the 12 conditions
(Science with MBBS, Science with Agric and Science with MS;
711
Technology with MBBS, Technology with Agric and Technology with MS;
Engineering with MBBS, Engineering with Agric and Engineering with
MS; Maths with MBBS, Maths with Agric and Maths with MS) in the study
are tabulated in the Descriptive Statistics. The Descriptive Statistics
information is used to partly answer the research questions.
Multivariate Tests
The null hypotheses of the main effects of STEM as well as of the
interaction effect of Stem*Discipline on the dependent variable, Srep
(Staff Recruitment Exercise Performance) could either be tested with
MANOVA (Multivariate Analysis of Variance) or with Univariate Analysis
of Variance. Results for four MANOVA tests (Pillai’s Trace, Wilk’s
Lambda, Hotelling’s Trace and Roy’s Largest Root) are given in this table.
The table displays the Value, the F, the Hypothesis df, the Error df, the
actual Sig. and the Partial Eta Squared; using each of the four tests to show
the effects of Stem and of Stem*Discipline interaction on the Staff
Recruitment Exercise Performance (Srep) of the applicants or participants.
The F for each of the tests is statistically significant (p < .05) in each case.
Meaning that (1) there is a significant effect of STEM on SREP, and (2)
there is a significant effect of STEM*Discipline interaction on SREP (p <
.0005 with substantially large effect size).
712
A few authors like Howell (2013; 2016) have noted that the Mauchly’s
Test of Sphericity is inaccurate and that it tends to be better to provide two
reports for the Within-Subjects ANOVA, one for Sphericity Assumed and
another for Sphericity not Assumed. Hence, double results of the Two-Way
Mixed (Between-Within) ANOVA shall be presented in this example
mainly for illustration sake. Ordinarily, when the sphericity assumption is
not met, the standard ANOVA F test given for Sphericity Assumed
inaccurately yields results that tend to reject a true null hypothesis slightly
more often than is true in the population; and therefore one of the alternate
F tests (Greenhouse-Geisser, Huynh-Feldt and Lower-bound) collectively
known as Epsilon, should rather be reported because these other tests adjust
the df to accurately accommodate the sphericity violation issue. Of the three
Epsilon or alternative F tests when sphericity assumption is violated or
when Sphericity Assumed F is to be bypassed due to slight elements of
inaccuracy in Mauchly’s Test, the best and most appropriately cautious
alternative F test to use is the Greenhouse-Geisser. The reason being that
while the Lower-bound F test makes adjustment that is overly cautious and
which might reject the null hypothesis a little too infrequently (i.e., fail to
reject the null hypothesis occasionally) due to its conservativeness, the
Huynh-Feldt F test makes adjustment that is slightly more liberal (rejecting
the null hypothesis a little more often than is warranted by the population
from which the representative sample was drawn).
713
and Mean Square for the Residual Term (Error) are also provided in the
table for each of the four tests (Sphericity Assumed, Greenhouse-Geisser,
Huynh-Feldt and Lower-bound). The Residual Term (Error) Mean Square is
the final denominator in each of the F tests.
The Tests of Within-Subjects Effects table provides the much-needed F
test as the ratio of the sources of variance, each represented as a Mean
Square (MS), which make up the General Linear Model Univariate
Analysis of Variance when repeated measures are taken. The F is the Mean
Square for Effect (STEM between the repeated measures) divided by the
Mean Square for Error (Residual Term) across the levels of STEM for any
of the four main effect F tests as well as for each of the four interaction
(Stem * Discipline) effect F tests (Sphericity Assumed, Greenhouse-
Geisser, Huynh-Feldt and Lower-bound).
In equation form, F for Sphericity Assumed regarding Stem main effect
is arrived at as Equation 10:1.
714
In both the Sphericity Assumed and Greenhouse-Geisser tests
scenarios, the main effect of Stem F is 3.120 which is as provided in the
table, and it is statistically significant as the Sig. (probability of obtaining
such rear result) is less than the chosen alpha of .05; and the Effect Size as
measured by Partial Eta Squared (ɳp2) is .052, a small effect size.
Therefore, the STEM intensive training has significant main effect on Srep,
the Staff Recruitment Exercise Performance: [F(3, 171) = 3.120, p < .05,
ɳp2 = .052 for Sphericity Assumed] and [F(2.848, 162.360) = 3.120, p <
.05, ɳp2 = .052 for Greenhouse-Geisser].
Recall that the determination of Effect Size in ANOVA models is
measured with Partial Eta Squired, and the Partial Eta Squired (ɳp2) is
interpreted as follows:
.000 - .009 = Trivial Effect Size
.010 - .059 = Small Effect Size
.060 - .139 = Medium Effect Size
.140 and above = Large Effect Size.
For the Stem * Discipline interaction effect in both Sphericity Assumed
and Greenhouse-Geisser tests situations, the F is 68.159 which is as
provided in the table, and it is significant statistically as the Sig. (p-value)
of .000 (read as less than .0005) is smaller than the chosen alpha of .05; it is
even less than .01 (if .01 were the alpha chosen). The Effect Size of the F
(68.159) as measured by Partial Eta Squared (ɳp2) is as large as .705.
Therefore, the null hypothesis that “there is no significant interaction effect
of Discipline and STEM training on the applicants’ staff recruitment
exercise performance” is rejected. STEM and Discipline have significant
interaction (Stem*Discipline) effect on Srep (Staff Recruitment Exercise
Performance): [F(6, 171) = 68.159, p < .05, ɳp2 = .705 (an
overwhelmingly large effect size) for Sphericity Assumed]; and
[F(5.697, 162.360) = 68.159, p < .05, ɳp2 = .705 (an overwhelmingly
large effect size) for Greenhouse-Geisser].
715
Tests of Within-Subjects Contrasts
The Tests of Within-Subjects Contrasts are not necessarily required in the
reporting and interpretation of Two-Way Mixed (Between-Within)
ANOVA. They may be required in some other statistical tests where they
play necessary roles. In fact, these tests are merely to indicate whether the
points where Linear line passes through are significantly straight; or
whether the points that the Quadratic U curves run are significant; or
whether the points significantly approximate Cubic or polynomial curve.
The Linearity, Quadratic and Cubic nature of the Stem and Stem *
Discipline interaction are depicted by the Profile Plots given towards the
end of the Output.
716
This table has given the Mean, Std. Error, and 95% Confidence Interval
Lower Bound and Upper Bound for each of the three levels of Discipline
(MBBS, Agric, and MS). The Std. Error for each of the levels is 2.583. The
Mean for MBBS is 38.750, for Agric is 45.375, and for MS is 41.125.
Pairwise Comparisons
Pairwise Comparisons of Discipline indicating the Mean Difference, Std.
Error, Sig., and Lower versus Upper Bounds at 95% Confidence Interval for
Difference on the basis of Bonferroni’s adjustment for multiple
comparisons show outright lack of significance for each of the pairwise
comparisons. That is, there is no pairwise comparison that has a significant
mean difference. IBM SPSS Statistics attaches an asterisk to any Mean
Difference (I – J) that is statistically significant. But there is no Mean
Difference with asterisk, showing that none of the pairwise Mean
Difference in Staff Recruitment Exercise Performance (Srep) is significant
with respect to Discipline.
Assessing the Pairwise Comparisons from the Sig. column, any Pairwise
Mean Difference that is significant must have a p-value (Sig.) that is lower
than the classically chosen alpha level of .05. But each of the Pairwise
Mean Differences is having Sig. that is higher than .05, showing therefore,
that none of the Pairwise Mean Differences is significant statistically.
Lastly, assessing from the 95% Confidence Interval for Difference
column, a Pairwise Mean Difference that is significant must have both the
Lower Bound and the Upper Bound either completely above zero or
completely below zero. But none of the Mean Differences meets this
condition. Therefore, all the Mean Differences in Srep with regard to
Discipline are a mere function of chance occurrence. Finally, for the fact
that the null hypothesis that “the applicants’ performance in the staff
recruitment exercise do not differ significantly with regards to their
discipline” is retained (not rejected) [F(2, 57) = 1.689, p > .05, ɳp2 = .056];
it is needless even showing any Post Hoc Multiple Comparison that border
on Mean Difference in staff recruitment exercise performance that is due to
Discipline.
Univariate Tests
717
This table clearly presents the F tests for the main effect of Discipline on
the dependent variable (Srep) on the basis of the linearly independent
pairwise comparisons among the estimated marginal means. The Sum of
Squares for Contrast (Discipline or Effect) is 450.625 and for Error is
7603.125. The df for Contrast is 2 and for Error is 57. The Mean Square for
Contrast is 225.313 and for Error is 133.388. The F is 1.689. That is:
The Sig. (p-value) is .194. Since the p > .05, the null hypothesis of no
significant effect of Discipline on Srep is retained (not rejected). The Effect
Size that is measured with Partial Eta Squared is .056.
Pairwise Comparisons
This table has displayed the 12 possible pairwise comparisons of Srep
means for the four levels of Stem. It has provided the Mean Difference (I-
J), Std. Error, Sig. and 95% Confidence Interval for Difference Lower
Bound and Upper Bound across the 12 pairs of the levels of Stem. For
718
instance, the Mean Difference between Science (1) and Technology (2) is
.667, Science (1) and Engineering (3) is -3.667, Science (1) and Maths (4)
is -.667. The Mean Difference between Technology (2) and Science (1) is
.667, Technology (2) and Engineering (3) is -4.333* (that is significant in
favour of Engineering), Technology (2) and Maths (4) is -1.333. The Mean
Difference between Engineering (3) and Science (1) is 3.667, Engineering
(3) and Technology (2) is 4.333* (significant in favour of Engineering),
Engineering (3) and Maths (4) is 3.000. The Mean Difference between
Maths (4) and Science (1) is .667, Maths (4) and Technology (2) is 1.333,
Maths (4) and Engineering (3) is -3.000.
The Mean Difference in each pair is got by subtracting the Mean for
Stem level (J) from the Mean for Stem level (I). Whenever the Stem level
(J) Mean is higher than the Stem level (I) Mean, a negative sign (-) is put as
prefix of the Mean Difference because the Mean of level J is subtracted
from the Mean of level I. If the Mean for Stem level J is smaller than the
Mean of Stem level I, there is no negative sign as prefix of the Mean
Difference as the mean J is subtracted from the mean I.
Any pair with a Mean Difference that is statistically significant has an
asterisk put for the Mean Difference to depict that the mean difference is
significant at the .05 alpha level. Of all the 12 pairwise comparisons, two
have the asterisk, showing that the Mean difference is significant. They are
Technology (2) and Engineering (3); and Engineering (3) and Technology
(2). The mean difference respectively of -4.333* and 4.333 is statistically
significant as the Sig. (.035) in each comparison is less than the chosen .05
alpha. Furthermore, for the mean difference of -4.333* the 95% Confidence
Interval for Difference Lower Bound (-8.464) and Upper Bound (-.203) are
completely below zero to show clearly that the mean difference is
statistically significant. For the mean difference of 4.333*, the 95%
Confidence Interval for Difference Lower Bound (.203) and Upper Bound
(8.464) are completely above zero to show clearly that the mean difference
is statistically significant.
All the other pairwise mean differences are not significant statistically as
(1), they do not have asterisk (*) assigned to them; (2), they have Sig. that
is greater than the chosen alpha level of .05 (p > .05); and (3), they have
95% Confidence Interval for Difference Lower Bound and Upper Bound
that do not completely fall either above zero or below zero in each
comparison.
719
Multivariate Tests
The Multivariate Tests are used to test the null hypotheses of the main
effect of Stem (Science, Technology, Engineering and Maths training) on
the dependent variable, Srep (Staff Recruitment Exercise Performance)
because the null hypothesis could either be tested with MANOVA
(Multivariate Analysis of Variance) or with Univariate Analysis of
Variance. Results for four MANOVA tests (Pillai’s Trace, Wilk’s Lambda,
Hotelling’s Trace and Roy’s Largest Root) are given in this table. The table
displays the Value, the F, the Hypothesis df, the Error df, the actual Sig. and
the Partial Eta Squared (measure of Effect Size); using each of the four tests
to show the effect of Stem intensive training on the Staff Recruitment
Exercise Performance (Srep) of the applicants who are the subjects or
participants in the investigation. The F (3.170) exact statistic for each of the
tests is statistically significant (p < .05) with a Partial Eta Squared of .147,
large effect size, in each test. Meaning that (1) there is a significant effect of
Stem on Srep [F(3.000, 55.000) = 3.170, p < .05, ɳp2 = .147]. That is,
Stem accounts for approximately 15% (exactly 14.7%) of the variance in
Srep. The MANOVA tests of the effect of Stem on Srep are each based on
the linearly independent pairwise comparisons among the estimated
marginal means, using exact statistic.
720
The interaction between MS (Management Science) and Science has
28.000 Mean and 3.319 Std. Error; MS and Technology interaction has
47.000 Mean and 2.806 Std. Error; MS and Engineering interaction has
30.000 Mean and 3.178 Std. Error, and MS with Maths interaction has
Mean and Std. Error of 59.500 and 2.871, respectively.
The Lower Bound and Upper Bound at 95% Confidence Interval for
each level of the interaction between Discipline and Stem can very easily
be read off as they are given in the table. Information in this table can be
used to answer the third research question with regard to interaction effect
of the between-subjects independent variable (Discipline) and the within-
subjects independent variable (Stem), Discipline * Stem, on the dependent
variable (Srep).
721
Profile Plots
The Profile Plots has graphically portrayed Estimated Marginal Means of
Srep (Staff Recruitment Exercise Performance) by Discipline (MBBS,
Agric, and MS) and Stem (1 = Science, 2 = Technology, 3 = Engineering,
and 4 = Maths); as well as the Discipline * Stem interaction. MBBS line
passes through Science at a mean of about 53, Technology at a mean of
about 31, Engineering at a mean of about 40 to Maths at a mean of
approximately 30. Applicants from Agric discipline have a Srep line chart
that passes through Science at a mean of about 41, Technology at a mean of
about 42, Engineering at a mean of about 63 to Maths at a mean of about
35. Applicants with MS (Management Science) discipline have a Srep line
chart that passes through Science at a mean of approximately 28,
Technology at the mean of about 48, Engineering at a mean of about 30 to
Maths at a mean of about 59. The linear, quadratic and cubic nature of the
Stem and Discipline as well as their complex interaction (Stem *
Discipline) effect on staff recruitment exercise performance (Srep) is
pictorially shown in the Profile Plots. The complex crisscrossing of the
lines demonstrates the substantially significant interaction effect of Stem
and Discipline on Srep. To disentangle the interaction effect, MANOVA
follow-up analysis was done, and it is interpreted in the next section. The
Profile Plots could also serve partly as pictorial illustration of answers to
the research questions.
722
the others. Aptly, the MANOVA was meant to reveal whether a significant
mean difference exists across all the levels of the Between-Subjects factor
(Discipline) for each Within-Subjects factor (Stem) level.
In other words, seven specific null hypotheses were tested with the
MANOVA, four in the first phase of the MANOVA tests and three in the
second dimension of the MANOVA tests. Apart from the seven major
MANOVA tests that breakdown the interaction effect of the Within-
Subjects independent variable (Stem) and the Between-Subjects
independent variable (Discipline) on the dependent variable (Srep), there
are five supplementary MANOVA tests that further clarify the interaction
effect. As it shall be seen very shortly, for each of the hypotheses tested
with the MANOVA follow-up for specification of the interaction effect of
Stem, the Within-Subjects factor, and Discipline, the Between-Subjects
factor, (i.e., the Stem*Discipline effect) on Srep (Staff Recruitment
Exercise Performance), the computed F is actually significant at as low as
.0005 alpha because the Sig of F in each of the tests is .000, read as less
than .0005. Bear in mind that the “Sig of F” in the MANOVA follow-up
output is the probability level (i.e., the p value) at which the specific result
(F) is statistically significant.
723
3251.67 MS, 14.76 F, and .000 Sig of F (read as less than .0005). The F of
14.76 is arrived at by dividing the MS Effect (i.e., for Discipline BY
MWITHIN STEM(1)) which is 3251.67 by the MS Error (i.e., for
WITHIN+RESIDUAL) which is 220.26. In equation form, the F is as
earlier given in Equation 10:1, but just note that the MS Effect which is the
denominator is made up of the WITHIN+RESIDUAL.
The Sig of F (.000) is less than the chosen .05 alpha. Therefore, the null
hypothesis of no significant mean difference across all the levels of the
Between-Subjects factor (Discipline) (MBBS, Agric and MS) for Science
level of Stem (the Within-Subjects factor) is rejected [F(2, 57) = 14.76, p
< .05].
724
Null hypothesis iii: There is no significant mean difference across the three
levels of Discipline (MBBS, Agric, MS) for the Engineering level of Stem.
Tests involving ‘MWITHIN STEM(3)’ Within-Subject Effect table of the
MANOVA results tested this null hypothesis.
The results show that the WITHIN+RESIDUAL (the Error term) has
11515.00 SS, 57 DF, and 202.02 MS. The MWITHIN STEM(3) has
118815.00 SS, 1 DF, 118815.00 MS, 588.14 F and .000 Sig of F. Crucially,
the Discipline BY MWITHIN STEM(3) has SS of 11370.00, DF of 2, MS
of 5685.00, F of 28.14, and .000 Sig of F. The Sig of F is lower than the
chosen alpha of .05. Therefore, the null hypothesis of no significant mean
difference across the three levels of Discipline (MBBS, Agric, and MS) for
the Engineering level of Stem is rejected [F(2, 57) = 28.14, p < .05].
725
Tests involving ‘STEM’ Within-Subject Effect
Null hypothesis v: There is no significant mean difference across the four
levels of the Within-Subjects factor (Stem) for MBBS level of Discipline
(the Between-Subjects factor).
The last MANOVA table in the Output 10.1, titled Tests involving
‘STEM’ Within-Subject Effect, provides the information for testing this
null hypothesis as well as for testing the next two null hypotheses. The table
shows that WITHIN+RESIDUAL has 12027.50 SS, 171 DF, and 70.34 MS.
MWITHIN DISCIPLINE(1) BY STEM has 7145.00 SS, 3 DF, 2381.67
MS, 33.86 F, and .000 Sig of F. The Sig of F is less than the .05 chosen
alpha level. Therefore, the null hypothesis is rejected [F(3, 171) = 33.86, p
< .05].
726
alpha at the appropriate degrees of freedom. Therefore, the null hypothesis
is rejected [F(3, 171) = 63.33, p < .05].
727
four tests are used to show the interaction effect of Discipline(3) which is
Management Science (MS) BY Stem intensive training on the Staff
Recruitment Exercise Performance (Srep).
728
p < .05, ɳp2 = .052 for Sphericity Assumed; and F(2.848, 162.4) = 3.120,
p < .05, ɳp2 = .052 for Greenhouse-Geisser. Pairwise comparison showed
significant Srep Mean Difference only between Engineering and
Technology in favour of Engineering: Mean Difference = 4.333, p < .05.
There was a significant interaction effect of Stem * Discipline on staff
recruitment exercise performance (Srep): F(6, 171) = 68.159, p < .05, ɳp2 =
.705 (an overwhelmingly large effect size) for Sphericity Assumed; and
F(5.697, 162.4) = 68.159, p < .05, ɳp2 = .705 (an overwhelmingly large
effect size) for Greenhouse-Geisser.
The MANOVA follow-up tests of the interaction effect of Stem *
Discipline on Srep has shown that:
1. The null hypothesis of no significant mean difference across all the
Between-Subjects factor (Discipline) levels (MBBS, Agric and MS)
for Science level of Stem (the Within-Subjects factor) is rejected
[F(2, 57) = 14.76, p < .05].
2. The null hypothesis of no significant mean difference across the three
levels of Discipline (the Between-Subjects factor) for the
Technology level of Stem (the Within-Subjects factor) is rejected
[F(2, 57) = 8.65, p < .05].
3. The null hypothesis of no significant mean difference across the three
levels of Discipline (MBBS, Agric, MS) for the Engineering level of
Stem is rejected [F(2, 57) = 28.14, p < .05].
4. The null hypothesis of no significant mean difference across the three
levels of Discipline (MBBS, Agric, MS) for the Engineering level of
Stem is rejected [F(2, 57) = 30.24, p < .05].
5. The null hypothesis of no significant mean difference across the four
levels of the Within-Subjects factor (Stem) for MBBS level of
Discipline (the Between-Subjects factor) is rejected [F(3, 171) =
33.86, p < .05].
6. The null hypothesis of no significant mean difference across all the
four levels of the Within-Subjects factor (Stem) for the Agric level
of Discipline (the Between-Subjects factor) is rejected [F(3, 171) =
42.24, p < .05].
7. The null hypothesis of no significant mean difference across all the
four levels of the Within-Subjects factor (Stem) for the MS level of
729
Discipline (the Between-Subjects factor) is rejected [F(3, 171) =
63.33, p < .05].
Step 6. Scroll up, down, left and right to view the output again in relation to
the interpretation offered for each of the tables.
Step 7. Save the output. Move cursor to the Toolbar and click the Save this
document icon to have the Save Output As dialog box. Type a
suitable name, Chapter 10 Output 1.spv, in the File name box (see Figure
10.33) and click Save to get the output securely saved for subsequent use or
reference whenever the need arises.
730
Figure 10.34 Save Syntax As… Chapter 10 Syntax 1.sps
May I emphasize that a Syntax file can only be saved in and retrieved
from the IBM SPSS Statistics Syntax Editor window; an Output
document can only be saved in and retrieved from the IBM SPSS Statistics
Viewer window. In like manner, a Data file can only be saved in and
retrieved from the IBM SPSS Statistics Data Editor.
Step 8. Print the output and the data files. Click the Print icon in the
Toolbar for Print dialog box to open. Ensure that All visible output is
checked, and click OK. With this click, one copy of the output gets printed.
Activate the IBM SPSS Statistics Data Editor containing the data, and
click the Print tool in the Toolbar, and click OK for a copy of the data to
also be printed. Similar actions can be taken to get a copy of the Syntax
printed after activation of the SPSS Statistics Syntax Editor that contains
the syntax in question. For more copies of the Output, Data or Syntax,
simple replication of the Print action up to the needed number suffices.
731
Step 9. Exit IBM SPSS Statistics by closing each of the SPSS window that
is open on the screen.
Step 10. Shut down the system by clicking Start Power Shut down.
732
The Syntax given here is made up of the main Two-Way Mixed
(Between-Within) ANOVA syntax and the MANOVA syntax for follow-up
tests when the Between factor and Within factor interaction effect is
significant statistically.
733
‘Seeing’ is believing. ‘Doing’ is knowing. There is no better way to
learn than by doing. There is no better way to prove that you have known
than by doing that which you have learnt for others to see, believe and be
convinced. What you have done is the greatest evidence that learning has
taken place. Needless much comments. So, let the analysis commence and
continue to most successful completion.
Switch on the computer
Window
734
Run
All
735
736
737
738
739
740
741
742
RESULTS OF MANOVA FOR FOLLOW-UP TESTS
743
744
745
746
INTERPRETATION OF OUTPUT 10.2 – EFFECTS
OF STEM AND DISCIPLINE
The Output 10.2 is as extensive as Output 10.1 with 28 sections that are
made up of 17 tables and 1 graph for the Two-Way Mixed (Between-
Within) ANOVA, and 10 sections for double MANOVA follow-up tests of
the significant interaction effect of the Between-Subject factor and the
Within-Subject factor on the dependent variable SREP. Again, not all the 28
sections that are necessarily required for reporting the findings. Carefully
compare the Output 10.1, when SPSS dialog boxes selection method was
used, with the Output 10.2, when the analysis was executed with SPSS
Syntax technique, to confirm that the outputs are exactly equivalent. The
interpretation offered for Output 10.1 depicts exactly how Output 10.2
should be interpreted. So, write out an interpretation of Output 10.2 in line
with the model that is provided for Output 10.1.
Adhere to the remaining protocols of data analysis with IBM SPSS
Statistics to eventually shut down the system.
View output
747
Print the output and the syntax
748
Chapter 11
ANALYSIS OF COVARIANCE
Overview
Analysis of Covariance, abbreviated as ANCOVA, is an extraordinarily
powerful inferential test for analysing data from an experimental
investigation to absolutely establish a cause-and-effect relationship, if it
does exist, between the manipulated independent variable and the
dependent variable when the influence of one or more covariates is
completely eliminated statistically. A covariate is an extraneous variable
that highly correlates with the dependent variable, and its influence, if not
held constant with ANCOVA, could easily be mistaken for the effect of the
manipulated independent variable on the dependent variable. This chapter
examines the uses of ANCOVA and the assumptions that underlie it; and
pragmatically demonstrates the execution of ANCOVA with IBM SPSS
Statistics syntax and dialog boxes point and click methods. With 53
illustrative SPSS screenshots and an excellent point-by-point easy to use
guide, every user (be it a total beginner or an experienced one) is sure to
attain expert knowledge in personal application of IBM SPSS Statistics in
performing ANCOVA, and in exceptional interpretation and reporting of the
results. ANOVA as a follow-up test for comparison of the F with that of the
ANCOVA is as well presented. Effect of lumosity training schedule on
recall of nonsense syllables, and lumosity training schedule effect on
intelligence quotient are used for exemplification.
749
quotient, Fluid intelligence, Interpretation of output, Summary of
output.
750
corresponding change either positively or negatively, in the latter. The most
straightforward approach for determination of causation (cause-and-effect
relationship) between two variables is application of the ANCOVA that
completely eliminates the effects of the covariates from the dependent
variable.
The opposite of Deterministic Covariation is Probabilistic Covariation.
Probabilistic Covariation is when the effect in the dependent variable is a
function of simultaneous manipulation of two or more independent
variables. In the case of probabilistic covariation, none of the two or more
independent variables can by itself alone be a necessary and sufficient cause
of the dependent variable; though all the independent variables are
collectively necessary for producing effect on the dependent variable
(Kpolovie, 2016; 2010). In other words, Probabilistic Covariation refers to a
causal relationship in which manipulation of each of the independent
variables increases the probability of change in the dependent variable, but
does not invariably produce the change. In probabilistic covariation, Two-
Way ANCOVA, Two-Way ANOVA, Three-Way ANOVA and or
Multiple regression is adopted for determination of causation or cause-
and-effect relationships between the independent variables and the
dependent variable. Where manipulation of one or more independent
variable(s) cause changes in two or more dependent variables, Multivariate
Analysis of Variance (MANOVA) is the suitable statistical tests to use.
Aptly put, the ANCOVA statistical test is adopted in investigation to
statistically control for factors that were not or cannot be controlled for
methodologically by methods such as random assignment or randomization.
Such factors may include sequencing or carry-over effect due to pre-test,
and some variables that definitely have linear relationships with the
dependent variable. Whenever there is probable evidence that a factor or
factors that cannot be controlled for methodologically might have an effect
on the dependent variable, ANCOVA should be used in the data analysis to
statistically control for, eliminate, hold constant, or covary such factor(s).
With the effect of the covariate adjusted for (held constant or partialed
out), the results of ANCOVA are interpreted in a manner that is similar to
that of any other analysis of variance. The only exception is that with
analysis of covariance, the adjusted means of the various groups instead of
the ordinary means are used. Adjusted means are the means of the various
groups after the effect of the covariate on them has been partialed out or
751
removed. There are five major situations that call for application of
ANCOVA. And the assumptions of ANCOVA are similar to those of other
parametric inferential statistics, and particularly to those of ANOVA.
Determination of change
The first major situation for use of analysis of covariance is when the
investigation is aimed at determination of change that has taken place as a
result of the introduction of experimental treatment between the pretest and
the posttest. Here, the pretest measure is taken and treated as the covariate.
752
ANCOVA reduces the error term
Analysis of covariance reduces the error term in ANOVA. The error term in
ANOVA which is decreased with the use of analysis of covariance is the
denominator that is also known as within group mean squares (MSWithin) or
mean square error (MSError). The smaller or lower the denominator, the
greater the probability of the calculated F to be statistically significant.
Analysis of covariance is able to reduce the error term because it adjusts for
(partials out) the effect of extraneous variables on the dependent variable.
The reduction of the error term accounts for why some times when data on
the dependent variable are analyzed with ANOVA, the difference is not
statistically significant; but when subjected to ANCOVA, significant
difference is found among the various groups that constitute the
independent variable. In other words, ANCOVA takes the portion of the
variance in the dependent variable that is not attributable to the independent
variable and removes it from the error term of an ordinary ANOVA. With
this adjustment, the subjects’ scores no longer vary because of the
covariate, and the remaining variance in the dependent variable is
accurately accounted for by only the independent variable, and can
therefore more easily attain statistical significance.
753
various groups is adjusted for, removed or partialed out to remain only the
mean differences that are totally attributable to the manipulation of the
independent variable. That is, one of the primary reasons for execution of
ANCOVA is to ascertain the true group differences on the dependent
variable that is purely due to the effect of only the independent variable
without any contribution of any confounding variable.
Analysis of covariance adjusts the group means on the dependent
variable downward where the covariate had unwanted positive influence. In
situations that the covariate had unwanted negative influence on the group
means, ANCOVA adjusts the mean differences upward to ensure that only
the effect of the independent variable determines the group means on the
dependent variable. Such adjustment of group means on the dependent
variable culminate in the increment of differences between group means
that is due to the independent variable alone and reduces the error term by
removing variance that is attributable to the covariate.
754
conditions is the ideal situation for application of analysis of covariance.
With effective random assignment of subjects into groups, the magnitude of
the covariate means for the different groups are equal; and any difference
between the groups, supposing the covariate was measured before the
treatments were applied, can only be due to chance occurrence. It is under
this condition that analysis of covariance best reduces the error term and
eliminates every bias in the means of the dependent variable by changing
group differences in the covariate to ensue true causation by the
independent variable alone.
With random assignment of subjects, analysis of covariance ensures that
any differences that exist on the pre-test are to be attributed to chance, and
the experimental and control groups will have equal means on the post-test
in the absence of a real treatment effect. Any difference in the post-test
must have been caused exclusively by the manipulation of the independent
variable.
755
the calculated ANCOVA F to be significant statistically (Field, 2013;
Cohen, 2008).
Homogeneity of regression
Homoscedasticity of regression is a very important assumption of
ANCOVA that the regression coefficients or slopes of the various groups
are all equal in the population. The assumption denotes that the correlation
between the covariate and the dependent variable is the same in each of the
groups when the variance of the dependent variable and the covariate are
the same or equal for all groups. This assumption also implies that there is
no interaction or at least no significant interaction between the covariate
and the independent variable because interaction can obscure or confound
the main effect of the independent variable on the dependent variable. Just
as homogeneity of variance is necessary for analysis of covariance, so is
homogeneity of the regression slopes or coefficients in the population.
756
EFFECT OF LUMOSITY TRAINING SCHEDULE
ON RECALL OF NONSENSE SYLLABLES
Kpolovie (2011) investigated the effects of Lumosity Training and Brain-
Boosting Food on learning. Findings indicated that training the brain with
Lumosity exercises or games significantly improved learning as the two
experimental groups that received lumosity training twice a day
significantly demonstrated learning more than the two groups that received
brain-boosting food/supplements which in turn, learned significantly better
than the two control groups in the study. As a follow-up, Kpolovie (2018)
studied the effect of Lumosity Training Schedule on the recall of nonsense
syllables. A random sample of 150 Tenth Grade students, each aged 15
years, was drawn for the investigation. They were randomized into five
groups that were randomly assigned different schedule of Lumosity
Training as experimental conditions. Each of the five groups had 30
students. The Lumosity Training Schedule for the groups were as follows:
Group 1 00 minute of lumosity training per day
Group 2 06 minutes of lumosity training per day
Group 3 12 minutes of lumosity training per day
Group 4 18 minutes of lumosity training per day
Group 5 24 minutes of lumosity training per day.
757
enough for the training to cause some changes in the subjects, if at all
lumosity training could positively or negatively affect the subjects’ memory
or recall ability; and the 60-day period was long enough for the subjects to
have forgotten the nonsense syllables as well as the pre-test that they took.
This measure was to reduce possibility of carry-over and test-wise effects.
After the last day of the experimental exposition, the entire 150 subjects
were presented the same 200 nonsense syllables for 20 minutes after which
the nonsense syllables were removed. Then, the same test on the nonsense
syllables that required the subjects to recall and write out as much of the
nonsense syllables as possible was administered to them. At the end of 30
minutes that the test lasted, the number of correct responses made by each
of the subjects was recorded as his/her post-treatment test (post-test) score.
The data collected for that investigation are used in this chapter as shown in
Table 11.1 to illustrate execution of Analysis of Covariance (ANCOVA).
758
759
760
* Nonsense syllables pre-test = NSPre
Nonsense syllables post-test = NSPost
Lumosity training schedule = LumosityTS (1 = 00minpd; 2 = 06minpd;
3 = 12minpd; 4 = 18minpd; 5 = 24minpd).
Research questions
Eleven (11) research questions (1 omnibus and 10 pairwise) were answered
at the end of the study with respect to the effect of Lumosity training
Schedule on recall of nonsense syllables when the Pre-test effect is
covaried:
1. What is the effect of Lumosity training schedule on the subjects’
recall of nonsense syllables when the influence of the pre-treatment
test has been covaried out as measured by their means and standard
deviations?
In other words, the omnibus research question is – do the five groups
differ in their means and standard deviations on the recall of
761
nonsense syllables when the influence of the Pre-test has been held
constant (controlled for)?
2. What is the difference between recall of nonsense syllables by those
who did not receive Lumosity training any day (00minpd) and those
who received 6 minutes Lumosity training per day (06minpd) when
the Pre-test is covaried?
3. What is the difference between recall of nonsense syllables by those
who did not receive Lumosity training any day (00minpd) and those
who received 12 minutes Lumosity training per day (12minpd) when
the Pre-test is covaried?
4. What is the difference between recall of nonsense syllables by those
who did not receive Lumosity training any day (00minpd) and those
who received 18 minutes Lumosity training per day (18minpd) when
the Pre-test is covaried?
5. What is the difference between recall of nonsense syllables by those
who did not receive Lumosity training any day (00minpd) and those
who received 24 minutes Lumosity training per day (24minpd) when
the Pre-test is covaried?
6. What is the difference between recall of nonsense syllables by those
who received 6 minutes Lumosity training per day (06minpd) and
those who received 12 minutes Lumosity training per day (12minpd)
when the Pre-test is covaried?
7. What is the difference between recall of nonsense syllables by those
who received 6 minutes Lumosity training per day (06minpd) and
those who received 18 minutes Lumosity training per day (18minpd)
when the Pre-test is covaried?
8. What is the difference between recall of nonsense syllables by those
who received 6 minutes Lumosity training per day (06minpd) and
those who received 24 minutes Lumosity training per day (24minpd)
when the Pre-test is covaried?
9. What is the difference between recall of nonsense syllables by those
who received 12 minutes Lumosity training per day (12minpd) and
those who received 18 minutes Lumosity training per day (18minpd)
when the Pre-test is covaried?
10. What is the difference between recall of nonsense syllables by those
who received 12 minutes Lumosity training per day (12minpd) and
762
those who received 24 minutes Lumosity training per day (24minpd)
when the Pre-test is covaried?
11. What is the difference between recall of nonsense syllables by those
who received 18 minutes Lumosity training per day (18minpd) and
those who received 24 minutes Lumosity training per day (24minpd)
when the Pre-test is covaried?
Alternate Hypotheses
Eleven alternate hypotheses (1 omnibus and 10 pairwise) were postulated in
the study as follows:
1. There is a significant effect of Lumosity training schedule on the
recall of nonsense syllables when the pre-treatment test (Pre-test)
influence has been covaried (held constant or controlled for).
2. There is a significant difference between recall of nonsense syllables
by those who did not receive Lumosity training any day (00minpd)
and those who received 6 minutes Lumosity training per day
(06minpd) when the Pre-test is covaried.
3. There is a significant difference between recall of nonsense syllables
by those who did not receive Lumosity training any day (00minpd)
and those who received 12 minutes Lumosity training per day
(12minpd) when the Pre-test is covaried.
4. There is a significant difference between recall of nonsense syllables
by those who did not receive Lumosity training any day (00minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is covaried.
5. There is a significant difference between recall of nonsense syllables
by those who did not receive Lumosity training any day (00minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is covaried.
6. There is a significant difference between recall of nonsense syllables
by those who received 6 minutes Lumosity training per day
(06minpd) and those who received 12 minutes Lumosity training per
day (12minpd) when the Pre-test is covaried.
7. There is a significant difference between recall of nonsense syllables
by those who received 6 minutes Lumosity training per day
763
(06minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is covaried.
8. There is a significant difference between recall of nonsense syllables
by those who received 6 minutes Lumosity training per day
(06minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is covaried.
9. There is a significant difference between recall of nonsense syllables
by those who received 12 minutes Lumosity training per day
(12minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is covaried.
10. There is a significant difference between recall of nonsense syllables
by those who received 12 minutes Lumosity training per day
(12minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is covaried.
11. There is a significant difference between recall of nonsense syllables
by those who received 18 minutes Lumosity training per day
(18minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is covaried.
Null Hypotheses
Similarly, eleven corresponding null hypotheses (1 omnibus and 10
pairwise) were postulated and tested at .05 in the study (Kpolovie, 2018) as
follows:
1. There is no significant effect of Lumosity training schedule on recall
of nonsense syllables when the Pre-treatment test (Pre-test) influence
has been covaried (held constant or controlled for).
2. No significant difference between recall of nonsense syllables by
those who did not receive Lumosity training any day (00minpd) and
those who received 6 minutes Lumosity training per day (06minpd)
when the Pre-test is covaried.
3. No significant difference between recall of nonsense syllables by
those who did not receive Lumosity training any day (00minpd) and
those who received 12 minutes Lumosity training per day (12minpd)
when the Pre-test is covaried.
4. No significant difference between recall of nonsense syllables by
those who did not receive Lumosity training any day (00minpd) and
764
those who received 18 minutes Lumosity training per day (18minpd)
when the Pre-test is covaried.
5. No significant difference between recall of nonsense syllables by
those who did not receive Lumosity training any day (00minpd) and
those who received 24 minutes Lumosity training per day (24minpd)
when the Pre-test is covaried.
6. No significant difference between recall of nonsense syllables by
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 12 minutes Lumosity training per day
(12minpd) when the Pre-test is covaried.
7. No significant difference between recall of nonsense syllables by
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is covaried.
8. No significant difference between recall of nonsense syllables by
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is covaried.
9. No significant difference between recall of nonsense syllables by
those who received 12 minutes Lumosity training per day (12minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is covaried.
10. No significant difference between recall of nonsense syllables by
those who received 12 minutes Lumosity training per day (12minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is covaried.
11. No significant difference between recall of nonsense syllables by
those who received 18 minutes Lumosity training per day (18minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is covaried.
765
Step 2. Launch IBM SPSS Statistics by clicking its icon on the computer,
click for instance.
Step 3. Enter the data carefully and accurately in the IBM SPSS Statistics
Data Editor just as shown in Table 11.1 with each subject’s three scores
(NSPre, NSPost and Lumosity) in one row consistently. On completion of
the data entry, move cursor to and click Variable View at the extreme left
bottom of the Data View. With this click, the Variable View of the IBM
SPSS Statistics Data Editor appears. In the first row under Name column,
type NSPre where is currently bearing VAR00001. In the second row under
Name column, type NSPost in place of VAR00002. In the third row under
Name column, type Lumosity in place of VAR00003. In the first, second
and third rows under Label column, type NSPre-test, NSPost-test and
LumosityTS, respectively. Under Measure column, select Scale for the
first row and second row. Nominal should be selected for the third row as
the independent variable, Lumosity, is a grouping or categorical variable.
Under Values column, click the ellipsis (small dotted blue box) for
the third row to have its dialog box in which you specify the value and label
for each group, level or category of the Lumosity Training Schedule. In
the Value Labels dialog box as shown in Figure 11.1, type 1 in the Value
box and 00minpd in the Label box. Then click Add for the value label to
be added into the big box at the right of Add.
766
In the box for Value, type 2, and in the box for Label, type 06minpd as
shown in Figure 11.2. Then, click Add for this value label to be added into
the big box at the right of Add.
Type 3 in the box for Value and 12minpd in the box for Label (see
Figure 11.3); and click Add for it to be added into the big box.
767
In the Value box, type 4. In the Label box, type 18minpd. After that,
click Add for this information to be added into the big box at the right of
, illustrated in Figure 11.4.
Type 5 in the Value box and 24minpd in the Label box as illustrated in
Figure 11.5; and click Add. A click on the Add moves the value label into
the big box.
768
Now, all the levels of the independent variable, Lumosity, have been
entered and moved into the big box as shown in Figure 11.6.
Click OK to return to the Variable View when all necessary naming and
specification of the variables have been done as illustrated in Figure 11.7.
To return to the Data View window of the IBM SPSS Statistics Data
Editor, move cursor to and click Data View at the extreme left bottom of
the Variable View. The Data View now has all the variables well labelled
as presented in Figure 11.8. The entire data as entered correctly and labelled
are ready for the analysis proper.
769
Figure 11.8 Data View completed
770
Figure 11.8 Continued
771
Figure 11.8 Continued
772
Figure 11.8 Continued
773
Figure 11.8 Continued
774
Figure 11.8 Continued
775
Figure 11.8 Continued
Step 4: Save the data. Click File, and click Save As. Alternatively, click the
Save this document icon in the Toolbar to have the Save Data As
dialog box displayed. Type the document’s name (Chapter 11 Table 1
Data.sav) in the File name box as shown in Figure 11.9 and click Save to
proceed.
776
Figure 11.9 Save Data As… Chapter 11 Table 1 Data.sav
777
On clicking the Univariate, its main dialog box appears as presented in
Figure 11.11. All the major ANCOVA statistical operations will be done in
or via the Univariate main dialog box. Do not get tired of returning back to
this main menu as you will be referred to it a number of times.
778
Figure 11.12 Univariate dialog box with variables moved to the right accordingly
779
Highlight Lumosity by clicking on it. Then, click the right-pointed
arrow that is facing Display Means for box so that it (Lumosity)
moves there as shown in Figure 11.14.
Figure 11.14 Univariate: Estimated Marginal Means with Lumosity moved to Display
Means for box
780
Figure 11.15 Univariate: Estimated Marginal Means with Compare main effects checked
Figure 11.16 Univariate: Estimated Marginal Means with Bonferroni checked for
Confidence interval adjustment
781
Figure 11.17 Univariate: Options
782
Figure 11.18 Univariate: Options with three statistical operations checked
783
Figure 11.19 Univariate: Profile Plots
784
Figure 11.20 Univariate: Profile Plots with Lumosity moved to Horizontal Axis
Then click Add for the Lumosity to be added in the plots. On clicking
the , Lumosity disappears from the Horizontal Axis box and rather
appears in the large box just below Plots as shown in Figure 11.21.
785
Figure 11.21 Univariate: Profile Plots with Lumosity Added
Click Continue to return to Univariate main dialog box when the entire
analysis has been done. Simply click OK in the Univariate main dialog box
and SPSS will rum the analysis and produce the output as presented in
Output 11.1.
786
For the purpose of clarification comparison, just very quickly perform
normal ANOVA (i.e., One-Way ANOVA) for the Nonsense Syllables Pre-
Test (NSPre) and Nonsense Syllables Post-Test (NSPost) so that you can
observe whether for the same NSPost scores, the F values arrived at with
ANCOVA is exactly the same as the F values arrived at with ANOVA.
Note that doing this One-Way ANOVA is not necessary, and it is not part
of the ANCOVA. In fact, the One-Way ANOVA does not have to be
reported and interpreted with the ANCOVA output as ANCOVA is on its
own a totally sufficient and completely adequate statistical test. I am just
performing the One-Way ANOVA here to further illustrate that indeed,
ANCOVA reduces the Within-Groups Variance (the Mean Square Within)
that serves as the final denominator, and increases the F value.
Activate the data saved as Chapter 11 Table 1 Data.sav as in Figure
11.8. Click Analyze Compare Means One-Way ANOVA (see
Figure 11.22).
787
Figure 11.23 One-Way ANOVA dialog box
Figure 11.24 One-Way ANOVA dialog box with variables moved to the right
788
Click Options pushbutton to have the One-Way ANOVA: Options
dialog box as in Figure 11.25.
789
Click Post Hoc pushbutton so that One-Way ANOVA: Post Hoc
Multiple Comparisons can also be done. The dialog box is shown in
Figure 11.27.
Figure 11.27 One-Way ANOVA: Post Hoc Multiple Comparisons dialog box
790
Figure 11.28 One-Way ANOVA: Post Hoc with Bonferroni checked
791
792
793
794
SECTION B OF OUTPUT 11.1 One-Way ANOVA
795
796
Means Plots
797
798
INTERPRETATION OF OUTPUT 11.1 – EFFECT OF
LUMOSITY TRAINING SCHEDULE
There are 13 segments in the Output 11.1 (8 for the ANCOVA proper and 5
for the Section B ANOVA follow-up test) on the effect of Lumosity training
schedule on recall of nonsense syllables. They are the Univariate Between-
Subjects Factors, Descriptive Statistics, Levene’s Test of Equality of Error
Variances, Tests of Between-Subjects Effects, Estimated Marginal Means
for LumosityTS, Pairwise Comparisons, Univariate Tests, and Profile Plots.
Others are the Oneway Descriptives, Test of Homogeneity of Variances,
ANOVA, Post Hoc Tests of Multiple Comparisons with Bonferroni, and
Means Plots. Each of these segments of the Output 11.1 is lucidly
interpreted herein.
799
Univariate Analysis of Variance Between-Subjects
Factors
This table merely lists the Value Label (00minpd, 06minpd, 12minpd,
18minpd and 24minpd) respectively for the five levels of LumosityTS
(1.00, 2.00, 3.00, 4.00 and 5.00). The N (number of cases) for each of the
five levels is 30.
Descriptive Statistics
The Descriptive Statistics table has presented the nonsense syllables post-
test (NSPost-test) Mean, Std. Deviation and N for each of the five levels of
LumosityTS. The table could serve as part of the answers to the research
questions. There is equal N of 30 for each level of the independent variable.
The recall of nonsense syllables has a Mean and Std. Deviation respectively
of 30.0000 and 4.82808 for 00minpd, 35.2667 and 5.07144 for 06minpd,
43.6000 and 5.64831 for 12minpd, 51.0667 and 5.18575 for 18minpd, and
54.1000 and 4.38925 for 24minpd. In all, the NSPost-test mean ranges from
30.0000 for 00minpd LumosityTS to 54.1000 for 24minpd LumosityTS.
The Total has 150 N, 42.8067 Mean and 10.42821 Std. Deviation.
800
summarised in this ANCOVA table is whether the five levels or treatment
conditions of Lumosity training schedule (LumosityTS) produce
significantly different scores on the dependent variable (recall of nonsense
syllables at the post-test). Information in the row in the table for the
independent variable (Lumosity) takes adequate care of the main concern of
the investigation. The Lumosity row has shown 12489.509 Type III Sum of
Squares, 4 df, 3122.377 Mean Square, 126.725 F, .000 (read as less than
.0005) Sig., and .779 Partial Eta Squared (ɳp2). The .000 Sig. is less than
the classically chosen alpha level of .05. Therefore, the omnibus null
hypothesis which is the first null hypothesis of the experimental study that
“there is no significant effect of Lumosity training schedule on recall of
nonsense syllables when the Pre-treatment test (Pre-test) influence has been
covaried (held constant)” is rejected. Thus, the alternate hypothesis that
indeed “there is a significant effect of Lumosity training schedule on recall
of nonsense syllables when the Pre-treatment test (Pre-test) influence has
been covaried (held constant)” is sustained. In other words, there is
preponderance of overwhelming evidence that the five LumosityTS groups
(00minpd, 06minpd, 12minpd, 18minpd, and 24minpd) significantly differ
in their means and standard deviations on the recall of nonsense syllables
after controlling for the effect of the Pre-test [F(4, 144) = 126.725, p < .05,
ƞp2 = .779, an overwhelmingly large effect size]. The Pairwise
Comparisons that will be discussed in a short while will reveal the specific
pairs of Lumosity training schedule that have statistically significant
difference.
801
Lumosity is 12489.509. It can equally be discerned from the Tests of
Between-Subjects Effects table that the Sum of Squares Error is 3548.017.
Thus, while the numerator of the Partial Eta squared is the SSLumosity which
is 12489.509, the denominator of the Partial Eta Squared is the SSLumosity
plus SSError, which is 12489.509 plus 3548.017 that is equal to 16037.526.
Therefore, the Partial Eta Squared is simply 12489.509 ÷ 16037.526 that is
equal to .779 as shown in the Lumosity row of the last column of the Tests
of Between-Subjects Effects table.
Perhaps, you may prefer explanation of the Effect Size in ANCOVA that
is measured with Partial Eta Squared (ƞp2) in equation form as presented
thus in Equation 11:1.
Since the Sum of Squares Effect in this case is actually the Sum of
Squares Lumosity, the Equation 11:1 can accordingly be specifically
rewritten as in Equation 11:2.
=
The Partial Eta Squared (ɳp2) of .779 is a marvellously Large Effect Size
in line with Cohen (1988) classification. This Effect Size (Partial Eta
Squared) denotes that Lumosity training schedule accounts for
approximately 78% or specifically, 77.9%, of variance in the Post-test recall
of nonsense syllables (NSPost-test) when the influence of the nonsense
syllables Pre-test (NSPre-test) is removed, eliminated, or controlled for.
I have in previous chapters discussed that the determination of Effect
Size in ANOVA models is measured with Partial Eta Squired, and that the
802
Partial Eta Squired (ɳp2) is interpreted as:
.000 - .009 = Trivial Effect Size
.010 - .059 = Small Effect Size
.060 - .139 = Medium Effect Size
.140 and above = Large Effect Size.
Another important finding, though not really the concern of this study, is
contained in the row that presents results for the covariate, the Nonsense
Syllables Pre-test (NSPre); showing 137.617 Type III Sum of Squares, 1 df,
137.617 Mean Square, 5.585 F, .019 Sig., and .037 Partial Eta Squared.
This row specifies whether there is a significant association between the
dependent variable, recall of nonsense syllables post-treatment test scores
(NSPost-test) and the covariate, recall of nonsense syllable pre-treatment
test scores (NSPre-test) when the effect of the independent variable
Lumosity training schedule (LumosityTS) is statistically controlled. The
Sig. for the covariate is .019 that is smaller than the .05 alpha; therefore,
association between the dependent variable and the covariate is statistically
significant when Lumosity training schedule is held under control. This also
implies that if the NSPre-test (Nonsense Syllables Pretest) were not
statistically eliminated, held constant or covaried out, it could have
significantly influenced the findings of the study (the pure effect of NSPost-
test on the dependent variable, recall of nonsense syllables).
In all, the Tests of Between-Subjects Effects has presented the Type III
Sum of Squares, df, Mean Square, F, Sig., and Partial Eta Squared (each of
these serving as a column) for each of the sources of variation, namely:
Corrected Model, Intercept, NSPre, and Lumosity. The F for each of these
sources is significant as the Sig. for it is lower than the .05 classical level of
significance. The table has equally displayed the Type III Sum of Squares
(SSTIII), df, and Mean Square for Error to be 3548.017, 144 and 24.639,
respectively. While the Total has 291065.000 SSTIII and 150 df, the
Corrected Total has 16203.393 SSTIII and 149 df, respectively.
803
has displayed the Nonsense Syllables Post-test (NSPost-test) Mean, Std.
Error, and 95% Confidence Interval Lower Bound and Upper Bound for the
five levels of Lumosity training schedule, LumosityTS, (00minpd,
06minpd, 12minpd, 18minpd, and 24minpd). For instance, 00minpd and
06minpd respectively have 30.010 and 35.296 Mean, .906 and .906 Std.
Error, 28.219 and 33.505 Lower Bound, and 31.802 and 37.088 Upper
Bound at 95% Confidence Interval. The Covariates appearing in the model
are evaluated at the following values: NSPre-test = 29.5867.
804
received 6 minutes Lumosity training per day (06minpd) when the Pre-test
is covaried in favour of the latter.
Null hypothesis 3 that “there is no significant difference between recall
of nonsense syllables by those who did not receive Lumosity training any
day (00minpd) and those who received 12 minutes Lumosity training per
day (12minpd) when the Pre-test is covaried” is rejected because the Mean
Difference of -13.568* has .000 Sig. (read as less than .0005) that is smaller
than .05 classically chosen alpha. Therefore, there is indeed a significant
difference between recall of nonsense syllables by those who did not
receive Lumosity training any day (00minpd) and those who received 12
minutes Lumosity training per day (12minpd) when the Pre-test is covaried
in favour of the latter.
Null hypothesis 4 that “there is no significant difference between recall
of nonsense syllables by those who did not receive Lumosity training any
day (00minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is covaried” is rejected. Reason for the
rejection is that the Mean Difference of -21.015* has .000 Sig. (read as less
than .0005) is smaller than the chosen alpha of .05. Therefore, there is a
significant difference between recall of nonsense syllables by those who did
not receive Lumosity training any day (00minpd) and those who received
18 minutes Lumosity training per day (18minpd) when the Pre-test is
covaried in favour of those who received 18 minutes Lumosity training per
day.
Null hypothesis 5 that “there is no significant difference between recall
of nonsense syllables by those who did not receive Lumosity training any
day (00minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is covaried” is rejected. Reason for the
rejection is that the Mean Difference of -24.113* has Sig. of .000 that is less
than the .05 chosen alpha. Therefore, there is a significant difference
between recall of nonsense syllables by those who did not receive Lumosity
training any day (00minpd) and those who received 24 minutes Lumosity
training per day (24minpd) when the Pre-test is covaried in favour of the
experimental group (those who received 24 minutes Lumosity training per
day).
Null hypothesis 6 that “there is no significant difference between recall
of nonsense syllables by those who received 6 minutes Lumosity training
per day (06minpd) and those who received 12 minutes Lumosity training
805
per day (12minpd) when the Pre-test is covaried” is rejected. The rejection
is because the Mean Difference of -8.282* has .000 Sig. that is less than .05
alpha. There is therefore a significant difference between recall of nonsense
syllables by those who received 6 minutes Lumosity training per day
(06minpd) and those who received 12 minutes Lumosity training per day
(12minpd) when the Pre-test is covaried. The difference is in favour of
12minpd (those who received 12 minutes per day Lumosity training).
Null hypothesis 7 that “there is no significant difference between recall
of nonsense syllables by those who received 6 minutes Lumosity training
per day (06minpd) and those who received 18 minutes Lumosity training
per day (18minpd) when the Pre-test is covaried” is rejected. Reason for
rejecting null hypothesis 7 is that the Mean Difference of -15.729* has Sig.
of .000 that is less than the classically chosen alpha of .05. There is
therefore a statistically significant difference between recall of nonsense
syllables by those who received 6 minutes Lumosity training per day
(06minpd) and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is covaried. The difference favours those who
received 18 minutes Lumosity training per day (18minpd).
Null hypothesis 8 that “there is no significant difference between recall
of nonsense syllables by those who received 6 minutes Lumosity training
per day (06minpd) and those who received 24 minutes Lumosity training
per day (24minpd) when the Pre-test is covaried” is rejected. This rejection
is because the Mean Difference of -18.827* has .000 Sig. (read as less than
.0005) that is lower than .05 classically chosen alpha. Therefore, there is a
significant difference between recall of nonsense syllables by those who
received 6 minutes Lumosity training per day (06minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the Pre-test
is covaried. The difference is in favour of those who received 24 minutes
Lumosity training per day (24minpd).
Null hypothesis 9 that “there is no significant difference between recall
of nonsense syllables by those who received 12 minutes Lumosity training
per day (12minpd) and those who received 18 minutes Lumosity training
per day (18minpd) when the Pre-test is covaried” is rejected. The reason for
rejecting it is that the Mean Difference of -7.447* has Sig. of .000 (read as
less than .0005) which is smaller than the chosen alpha of .05. Therefore,
there is indeed a statistically significant difference between recall of
nonsense syllables by those who received 12 minutes Lumosity training per
806
day (12minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is covaried. The difference is in favour of
the latter (those who had 18minpd Lumosity training).
Null hypothesis 10 that “there is no significant difference between recall
of nonsense syllables by those who received 12 minutes Lumosity training
per day (12minpd) and those who received 24 minutes Lumosity training
per day (24minpd) when the Pre-test is covaried” is rejected. This rejection
is because the Mean Difference of -10.545* has Sig. of .000 which is less
than the classically chosen alpha of .05. Therefore, there is a significant
difference between recall of nonsense syllables by those who received 12
minutes Lumosity training per day (12minpd) and those who received 24
minutes Lumosity training per day (24minpd) when the Pre-test is covaried.
The significant mean difference is in favour of those who had 24 minutes
Lumosity training per day.
Null hypothesis 11 that “there is no significant difference between recall
of nonsense syllables by those who received 18 minutes Lumosity training
per day (18minpd) and those who received 24 minutes Lumosity training
per day (24minpd) when the Pre-test is covaried” is retained (failed to
reject). Reason for the failure to reject null hypothesis 11 is that the Mean
Difference of -3.098 has Sig. of .169 that is greater than the classically
chosen alpha of .05. Thus, the alternate hypothesis that “there is a
significant difference between recall of nonsense syllables by those who
receive 18 minutes Lumosity training per day (18minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the Pre-test
is covaried” is discarded. What seems to be like a mean difference between
the two levels of Lumosity training (18 minutes per day and 24 minutes per
day) with regard to the post-test scores on recall of nonsense syllables when
the pre-test is controlled for, is merely a function of chance that is not
consistent; and cannot qualify for a significant difference.
In addition to testing ten of the null hypotheses, information in the
Pairwise Comparison of NSPost-test section of the Output 11.1 can also be
used to answer the research questions. The Pairwise Mean Difference in
each situation, is an answer to the corresponding research question as
outlined in Table 11.1a. Each significant difference is in favour of the group
with more minutes of exposure to LumosityTS per day.
807
Table 11.1a Answers to the research questions
No Research Question Answer Hypothesis
Decision
i. What is the effect of Lumosity 00minpd Significant
training schedule on the subjects’ M=30.010 F(4, 144) =
recall of nonsense syllables when the SD=4.82808 126.725, p
influence of the pre-treatment test 06minpd < .05, ɳp2 =
has been covaried out as measured M=35.296 .779
by their means and standard SD=5.07144
deviations? In other words, the 12minpd
omnibus research question is – do M=43.578
the five groups differ in their means SD=5.64831
and standard deviations on the recall 18minpd
of nonsense syllables when the M=51.025
influence of the Pre-test has been SD=5.18575
held constant? 24minpd
M=54.123
SD=4.38925
ii. What is the difference between recall -5.286 Significant
of nonsense syllables by those who p < .05
did not receive Lumosity training
any day (00minpd) and those who
received 6 minutes Lumosity
training per day (06minpd) when the
Pre-test is covaried?
iii. What is the difference between recall -13.568 Significant
of nonsense syllables by those who p < .05
did not receive Lumosity training
any day (00minpd) and those who
received 12 minutes Lumosity
training per day (12minpd) when the
Pre-test is covaried?
iv. What is the difference between recall -21.015 Significant
of nonsense syllables by those who p < .05
did not receive Lumosity training
808
any day (00minpd) and those who
received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is covaried?
v. What is the difference between recall -24.113 Significant
of nonsense syllables by those who p < .05
did not receive Lumosity training
any day (00minpd) and those who
received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is covaried?
vi. What is the difference between recall -8.282 Significant
of nonsense syllables by those who p < .05
received 6 minutes Lumosity
training per day (06minpd) and those
who received 12 minutes Lumosity
training per day (12minpd) when the
Pre-test is covaried?
vii. What is the difference between recall -15.729 Significant
of nonsense syllables by those who p < .05
received 6 minutes Lumosity
training per day (06minpd) and those
who received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is covaried?
viii. What is the difference between recall -18.827 Significant
of nonsense syllables by those who p < .05
received 6 minutes Lumosity
training per day (06minpd) and those
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is covaried?
ix. What is the difference between recall -7.447 Significant
of nonsense syllables by those who p < .05
received 12 minutes Lumosity
809
training per day (12minpd) and those
who received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is covaried?
x. What is the difference between recall -10.545 Significant
of nonsense syllables by those who p < .05
received 12 minutes Lumosity
training per day (12minpd) and those
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is covaried?
xi. What is the difference between recall -3.098 Not
of nonsense syllables by those who significant
received 18 minutes Lumosity p > .05
training per day (18minpd) and those
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is covaried?
Univariate Tests
The Univariate Tests table presents the NSPost-test Sum of Squares, df,
Mean Square, F, Sig., and Partial Eta Squared for Contrast and Error. The F
is 126.377 with .000 Sig. (read as less than .0005), and is statistically
significant as p < .05; and the Partial Eta Squared is .779. That is, tenability
of the omnibus null hypothesis tested by the Univariate Tests is rejected
[F(4, 144) = 126.725, p < .05, ɳp2 = .779]. This F indicates the effect of
LumosityTS on the basis of linearly independent pairwise comparisons
among the estimated marginal means.
Profile Plots
The Profile Plots serve as pictorial illustration of answers to the research
questions. The Plots indicate the Estimated Marginal Means of Post-test
recall of Nonsense Syllables as a function of the Lumosity training schedule
of 00minpd, 06minpd, 12minpd, 18minpd, and 24minpd. Consistent
increase in the recall of Nonsense Syllables (from a mean of 30.000 for
810
00minpd to a mean of 54.123 for 24minpd) in direct accordance with the
increase in the length of exposure to the experimental treatment condition
(Lumosity Training Schedule) can most easily be depicted as pictorially
captured by the Profile Plots.
Oneway Descriptives
The Oneway Descriptives table is a display of the N, Mean, Std. Deviation,
Std. Error, 95% Confidence Interval for Mean Lower Bound and Upper
Bound, the Minimum and Maximum of recall of nonsense syllables for
each of the LumosityTS levels (00minpd, 06minpd, 12minpd, 18minpd, and
24minpd) at the Pre-test (NSPre-test) and at the Post-test (NSPost-test). For
instance, the N is 30 for each of the LumosityTS levels both at NSPre-test
and NSPost-test. At the NSPre-test, the Mean is 29.5333 for 00minpd,
29.4333 for 06minpd, 29.7000 for 12minpd, 29.8000 for 18minpd, and
29.4667 for 24minpd; the Total has 150 N and 29.4667 Mean. That is, at the
NSPre-test the recall of nonsense syllables means ranged only from 29.4333
to 29.8000. At the NSPost-test, the recall of nonsense syllables has the
Mean of 30.0000 for 00minpd, 35.2667 for 06minpd, 43.6000 for 12minpd,
51.0667 for 18minpd, and 54.1000 for 24minpd. The Total has 150 N and
42.8067 mean. That is, at the NSPost-test the recall of nonsense syllables
means ranged from 30.0000 and consistently increased to 54.1000. While
mean for Total during NSPre-test is 29.5867, the mean for Total during
NSPost-test is 42.8067. All the other values in the table can easily be read
and interpreted similarly.
811
ANOVA
The ANOVA table has shown that for NSPre-test, the Sum of Squares
Between Groups is 2.973, df Between Groups is 4, and the Mean Square
Between Groups is .743. The Within Groups at NSPre-test has 3693.400
Sum of Squares, 145 df, and 25.472 Mean Square. The Total Sum of
Squares and df are respectively 3696.373 and 149. The F is .029 with .998
Sig. Since the Sig. of .998 is greater than the classically chosen .05 alpha,
the omnibus null hypothesis of no significant effect of Lumosity training
schedule on recall of nonsense syllables at the Pre-test is retained (not
rejected).
The Effect Size, measured with Partial Eta Squared is .001 that is got by
dividing the Sum of Squares Between Groups (SSB) by the Total Sum of
Squares (SST). In simple One-Way ANOVA, Sum of Squares Between
Groups is equivalent to Sum of Squares Effect in higher ANOVA models;
and the Sum of Squares Within Groups in One-Way ANOVA is an
equivalent of Sum of Squares Error in higher ANOVA models.
Furthermore, Sum of Squares Total in simple One-Way ANOVA is equal to
the Sum of Squares Between Groups plus the Sum of Squares Within
Groups. That is why, the Partial Eta Squared (ƞp2) equation in the simple
One-Way ANOVA can merely be given as in Equation 11:3.
812
of Squares and df are 16203.393 and 149, respectively. At the NSPost-test,
the F is 123.118 with .000 Sig. (read as less than .0005). The Sig. (.000) is
less than the classically chosen alpha of .05. Therefore, the omnibus null
hypothesis of no significant effect of Lumosity training schedule on recall
of nonsense syllables at the Post-test is rejected. You could recall that in the
ANCOVA, the F was 126.725 that is greater than the F of 123.118 in the
ANOVA, a F difference of 3.607 in favour of the ANCOVA. Accordingly,
the Error Mean Square in the ANCOVA is 24.639, while in the ANOVA,
the Within Groups Mean Square is 25.418.
813
asterisk because for each Mean Difference, the Sig. is greater than .05
alpha.
At the NSPost-test, all the Pairwise Comparisons, except one (18minpd
and 24minpd), are statistically significant because the Sig. in each of the
cases is less than the classically chosen .05 alpha. Consequently, there is
asterisk attached to each of the Mean Differences to show that the mean
difference is significant at the .05 level. The difference in each case is in
favour of the Lumosity TS that has higher mean, the group with the
lumosity training schedule which invariably received more minutes of the
experimental exposure.
Means Plots
There are two Means Plots, one for NSPre-test and the other for NSPost-
test. The first Means Plots is for the NSPre-test and it shows that for each of
the five levels of LumosityTS, the mean is approximately within 29.43 and
29.80. The second Means Plots clearly show remarkable increment
(depicting significance) from the mean of 30.00 for 00minpd through mean
of approximately 35.00 for 06minpd, 44.00 for 12minpd, to 51.00 for
18minpd. The mean for 24minpd LumosityTS is approximately 54.00,
higher than that of 18minpd, though not significant.
814
00minpd and 06minpd [MD = -5.286, p < .05]; 00minpd and 12minpd [MD
= -13.568, p < .05]; 00minpd and 18minpd [MD = -21.015, p < .05];
00minpd and 24minpd [MD = -24.113, p < .05]. There were also
statistically significant Mean Difference (MD) between 06minpd and
12minpd [MD = -8.282, p < .05]; 06minpd and 18minpd [MD = -15.729, p
< .05]; 06minpd and 24minpd [MD = -18.827, p < .05]; 12minpd and
18minpd [MD = -7.447, p < .05]; and 12minpd and 24minpd [MD =
-10.545, p < .05]. In each of the Pairwise Comparisons, the Mean
Difference was in favour of the group that had higher minutes of Lumosity
training per day. The Mean Difference (MD) between 18minpd and
24minpd was not significant [MD = -3.098, p > .05]. Profile Plots was
drawn to pictorially illustrate the recall of nonsense syllables Means across
the five levels of LumosityTS.
One-Way ANOVA was done as follow-up to ascertain the results
supposing the NSPre-test was not controlled for. Results showed that effect
of LumosityTS on the NSPre-test is not significant at all [F(4, 145) = .029,
p > .05, ɳp2 = .001]. Post Hoc Multiple Comparisons confirmed that none of
the pairwise comparisons differ significantly as the Sig. in each case is
greater than .05 alpha. Means Plots for the nonsense syllables recall during
the Pre-test showed means that ranged from approximately 29.43 to 29.80.
But the effect of LumosityTS on the NSPost-test was significant
statistically [F(4, 145) = 123.118, p < .05, ɳp2 = .773]. Post Hoc Multiple
Comparisons showed significant pairwise Mean Difference (MD) between
00minpd and 06minpd [MD = -5.26667, p < .05]; 00minpd and 12minpd
[MD = -13.60000, p < .05]; and 00minpd and 18minpd [MD = -21.06667, p
< .05]; 00minpd and 24minpd [MD = -24.10000, p < .05]. There were also
statistically significant Mean Difference (MD) between 06minpd and
12minpd [MD = -8.33333, p < .05]; 06minpd and 18minpd [MD =
-15.80000, p < .05]; 06minpd and 24minpd [MD = -18.83333, p < .05];
12minpd and 18minpd [MD = -7.46667, p < .05]; and 12minpd and
24minpd [MD = -10.50000, p < .05]. In each of the Pairwise Comparisons,
the Mean Difference was in favour of the group that was exposed to more
minutes of Lumosity training per day. The Mean Difference (MD) between
18minpd and 24minpd was not significant [MD = -3.03333, p > .05].
Means Plots was drawn to pictorially illustrate the recall of nonsense
syllables Means across the five levels of LumosityTS. It showed that the
mean consistently increased from one level of LumosityTS (00minpd with
815
mean of 30.00 as the lowest) to 24minpd with a mean of approximately
54.00 as the highest.
Step 7. Save the output. Click the Save this document icon in the Toolbar
to have the Save Output As dialog box. In it, type the suitable name,
Chapter 11 Output 1.spv, in the File name box; and click Save.
Step 8. Print the output and then the data. In each case, be on the window
and click the Print tool in the Toolbar. Then, click OK to have the
document printed.
Step 9. Exit IBM SPSS Statistics by closing each of SPSS window that is
open on the computer.
Step 10. Shut down the system by clicking Start, Power and Shut down
accordingly.
816
Switch the computer on and let it boot fully. Launch IBM SPSS
Statistics by clicking its logo on your system. Then, wait for it to fully
get ready. When due, close the Welcome to IBM SPSS Statistics dialog
box that is superimposed on the SPSS Data Editor.
Retrieve the Chapter 11 Table 1 Data.sav. To do this, click Open this
document tool in the Toolbar. This will produce the Open Data dialog box.
In it, look for the file name, Chapter 11 Table 1 Data.sav, click it and
press the Enter key on the Keyboard. This action opens the data as shown
Figure 11.8.
Click Window and from its pulldown menu, click Go to Designated
Syntax Window. With this, the IBM SPSS Statistics Syntax Editor
opens. Correctly enter the Command given in Syntax 11.1. Use the help
offered by SPSS in prompting you while typing the Syntax. When done, the
Syntax Editor will look like Figure 11.29.
817
Figure 11.29 Chapter 11 Syntax 1.sps
Highlight the entire Syntax and click Run Selection icon . With
this click, IBM SPSS Statistics completely performs the analysis and
produces the results in the Viewer window as presented in Output 11.2.
Alternatively, do not highlight the Syntax; simply click Run and in its
pulldown menu, click All and the whole analysis gets done and the results
produced in the IBM SPSS Statistics Viewer as illustrated in Output 11.2.
818
819
820
821
SECTION B OF OUTPUT 11.1 One-Way ANOVA
822
823
Means Plots
824
825
INTERPRETATION OF OUTPUT 11.2
The Output 11.1 is the same with Output 11.2; though while the former was
arrived at through dialog boxes point and click method, the latter was a
product of IBM SPSS Statistics Syntax application on the data. Therefore,
the same interpretation for Output 11.1 goes for the Output 11.2. Any
hypothesis that is rejected, using Output 11.1 is also rejected with the use of
Output 11.2. Pairwise Multiple Comparisons that are significant in Output
11.1 are the ones that are statistically significant in Output 11.2.
Follow the remaining SPSS data analysis protocols to eventually shut
down your system.
View output
826
Save the output as Chapter 11 Output 2.spv and the Syntax as Chapter
11 Syntax 1.sps
827
18 minutes Lumosity training per day for the 30 subjects in Group 4
(18minpd)
24 minutes Lumosity training per day for the 30 subjects in Group 5
(24minpd).
The Intelligence Quotients of the subjects at pre-test and post-test are as
tabulated in Table 11.2.
828
829
Research questions
One omnibus research question and ten pairwise research questions were
answered at the end of the study with respect to the effect of Lumosity
Training Schedule (LumosityTS) on Intelligence Quotient, IQ, when the IQ
830
Pre-test effect is controlled for or held constant. The eleven research
questions are as enumerated.
1. What is the effect of Lumosity Training Schedule on the subjects’
Intelligence Quotient (IQ) when the influence of the Pre-treatment IQ
test has been controlled for or held constant as measured by their
means and standard deviations? (That is: Do the five Groups differ in
their IQ means and standard deviations when the influence of the IQ
pre-test has been held constant?)
2. What is the difference between the IQ of those who did not receive
Lumosity training any day (00minpd) and those who received 6
minutes Lumosity training per day (06minpd) when the Pre-test is
controlled for?
3. What is the difference between the IQ of those who did not receive
Lumosity training any day (00minpd) and those who received 12
minutes Lumosity training per day (12minpd) when the Pre-test is
controlled for?
4. What is the difference between the IQ of those who did not receive
Lumosity training any day (00minpd) and those who received 18
minutes Lumosity training per day (18minpd) when the Pre-test is
controlled for?
5. What is the difference between the IQ of those who did not receive
Lumosity training any day (00minpd) and those who received 24
minutes Lumosity training per day (24minpd) when the Pre-test is
controlled for?
6. What is the difference between the IQ of those who received 6
minutes Lumosity training per day (06minpd) and those who
received 12 minutes Lumosity training per day (12minpd) when the
Pre-test is controlled for?
7. What is the difference between the IQ of those who received 6
minutes Lumosity training per day (06minpd) and those who
received 18 minutes Lumosity training per day (18minpd) when the
Pre-test is controlled for?
8. What is the difference between the IQ of those who received 6
minutes Lumosity training per day (06minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the
Pre-test is controlled for?
831
9. What is the difference between the IQ of those who received 12
minutes Lumosity training per day (12minpd) and those who
received 18 minutes Lumosity training per day (18minpd) when the
Pre-test is controlled for?
10. What is the difference between the IQ of those who received 12
minutes Lumosity training per day (12minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the
Pre-test is controlled for?
11. What is the difference between the IQ of those who received 18
minutes Lumosity training per day (18minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the
Pre-test is controlled for?
Alternate Hypothesis
Eleven alternate hypotheses (1 omnibus and 10 pairwise) were postulated in
the study as itemised.
1. There is a significant effect of Lumosity training schedule on the
subjects’ Intelligence Quotient (IQ) when the pre-treatment test (Pre-
test) influence has been controlled for (held constant).
2. There is a significant difference between the Intelligence Quotient of
those who did not receive Lumosity training any day (00minpd) and
those who received 6 minutes Lumosity training per day (06minpd)
when the Pre-test is controlled for.
3. There is a significant difference between the Intelligence Quotient of
those who did not receive Lumosity training any day (00minpd) and
those who received 12 minutes Lumosity training per day (12minpd)
when the Pre-test is controlled for.
4. There is a significant difference between the Intelligence Quotient of
those who did not receive Lumosity training any day (00minpd) and
those who received 18 minutes Lumosity training per day (18minpd)
when the Pre-test is controlled for.
5. There is a significant difference between the Intelligence Quotient of
those who did not receive Lumosity training any day (00minpd) and
those who received 24 minutes Lumosity training per day (24minpd)
when the Pre-test is controlled for.
832
6. There is a significant difference between the Intelligence Quotient of
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 12 minutes Lumosity training per day
(12minpd) when the Pre-test is controlled for.
7. There is a significant difference between the Intelligence Quotient of
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is controlled for.
8. There is a significant difference between the Intelligence Quotient of
those who received 6 minutes Lumosity training per day (06minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is controlled for.
9. There is a significant difference between the Intelligence Quotient of
those who received 12 minutes Lumosity training per day (12minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is controlled for.
10. There is a significant difference between the Intelligence Quotient of
those who received 12 minutes Lumosity training per day (12minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is controlled for.
11. There is a significant difference between the Intelligence Quotient of
those who received 18 minutes Lumosity training per day (18minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is controlled for.
Null Hypothesis
Similarly, eleven corresponding null hypotheses (1 omnibus and 10
pairwise) were postulated as follows and tested at .05 alpha in the study
(Kpolovie, 2018; 2018a).
1. There is no significant effect of Lumosity training schedule on the
subjects’ Intelligence Quotient (IQ) when the pre-treatment test (Pre-
test) influence has been controlled for (held constant).
2. There is no significant difference between the Intelligence Quotient
of those who did not receive Lumosity training any day (00minpd)
and those who received 6 minutes Lumosity training per day
(06minpd) when the Pre-test is controlled for.
833
3. There is no significant difference between the Intelligence Quotient
of those who did not receive Lumosity training any day (00minpd)
and those who received 12 minutes Lumosity training per day
(12minpd) when the Pre-test is controlled for.
4. There is no significant difference between the Intelligence Quotient
of those who did not receive Lumosity training any day (00minpd)
and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is controlled for.
5. There is no significant difference between the Intelligence Quotient
of those who did not receive Lumosity training any day (00minpd)
and those who received 24 minutes Lumosity training per day
(24minpd) when the Pre-test is controlled for.
6. There is no significant difference between the Intelligence Quotient
of those who received 6 minutes Lumosity training per day
(06minpd) and those who received 12 minutes Lumosity training per
day (12minpd) when the Pre-test is controlled for.
7. There is no significant difference between the Intelligence Quotient
of those who received 6 minutes Lumosity training per day
(06minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is controlled for.
8. There is no significant difference between the Intelligence Quotient
of those who received 6 minutes Lumosity training per day
(06minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is controlled for.
9. There is no significant difference between the Intelligence Quotient
of those who received 12 minutes Lumosity training per day
(12minpd) and those who received 18 minutes Lumosity training per
day (18minpd) when the Pre-test is controlled for.
10. There is no significant difference between the Intelligence Quotient
of those who received 12 minutes Lumosity training per day
(12minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is controlled for.
11. There is no significant difference between the Intelligence Quotient
of those who received 18 minutes Lumosity training per day
(18minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is controlled for.
834
ANCOVA WITH SPSS DIALOG BOXES POINT
AND CLICK
Step 1. Boot the computer.
Step 3. Enter the data in Table 11.2 carefully and accurately. Follow all the
procedures that were used to enter, name and label the recall of Nonsense
Syllables scores to enter, name and label the dependent variable
(Intelligence Quotient Post-Test), the independent variable (Lumosity
Training Schedule), and the covariate (Intelligence Quotient Pre-Test).
Ensure that the three scores of each participant consistently constitute one
row in the SPSS Data Editor exactly as shown in Table 11.2. In the SPSS
Data View, respectively Name and Label the Intelligence Quotient Pre-
Test as IQPre and Pre-Test IQ, the Intelligence Quotient Post-Test as
IQPost and Post-test IQ, and the Lumosity Training Schedule as
Lumosity and LumosityTS. After entry of the data in the Data View of the
IBM SPSS Statistics Data Editor and the naming and labelling of the
variables in the Variable View, the Variable View of the IBM SPSS
Statistics Data Editor will look like Figure 11.30.
Figure 11.30 Variable View with the variables named and labelled.
When you return to the Data View of the SPSS Data Editor by clicking
Data View at the extreme bottom left of the Variable View, the entire data
as entered and named will look like Figure 11.31.
835
Figure 11.31 Data Editor with the data entered, named and labelled
836
Figure 11.31 Continued
837
Figure 11.31 Continued
838
Figure 11.31 Continued
839
Figure 11.31 Continued
840
Figure 11.31 Continued
841
Figure 11.31 Continued
Step 4. Save the data as Chapter 11 Table 2 Data.sav by clicking the Save
this document icon in the Toolbar and typing the document name in
the File name box of the Save Data As, and finally clicking Save to get the
data saved for subsequent use when the need arises.
842
Figure 11.32 Save Data As… Chapter 11 Table 2 Data.sav
843
A click on the Univariate produces the main Univariate dialog box,
illustrated in Figure 11.34, in which the requisite statistical operations will
be performed.
Click to highlight Post-test IQ in the left box and move it into the
Dependent Variable box at the right by clicking the right-pointed arrow
that is directly facing the Dependent Variable panel. Select
LumosityTS in the left box and move it into the Fixed Factor(s) box at the
right by clicking the right-pointed arrow that is facing the Fixed
Factor(s) box directly. Select Pre-test IQ in the left box and move it into
the Covariate(s) field at the right. The movement is like in other situations
that is done by clicking the right-pointed arrow that is facing the
Covariate(s) box at the right. When done, the Univariate dialog box will
look as exemplified in Figure 11.35.
844
Figure 11.35 Univariate dialog box with variables moved right
845
Select Lumosity at the left in Factor(s) and Factor Interactions box
and transfer it with the help of the right-pointed arrow into the Display
Means for panel at the right as shown in Figure 11.37. The movement
activates Compare main effects.
Figure 11.37 Univariate: Estimated Marginal Means with the independent variable
(Lumosity) moved to the right.
846
Figure 11.38 Univariate: Estimated Marginal Means with compare main effects checked
847
Figure 11.40 Univariate: Estimated Marginal Means with Bonferroni selected
Click Continue to return to the Univariate main dialog box and select
Options pushbutton to have the Univariate: Options dialog box, illustrated
in Figure 11.41.
848
Figure 11.41 Univariate: Options
849
Figure 11.42 Univariate: Options with the necessary operations checked
Click Continue to return to the main Univariate dialog box. In it, select
Plots pushbutton for Univariate: Profile Plots dialog box to appear as
given in Figure 11.43.
850
Figure 11.43 Univariate: Profile Plots
851
Figure 11.44 Univariate: Profile Plots with Lumosity moved accordingly
852
Figure 11.45 Univariate: Profile Plots with Lumosity Added
Get back to the Univariate main dialog box by clicking Continue. Now,
all the required statistical operations have been done. It is just ready for the
results to be displayed. So finally, move cursor to, and click OK for IBM
SPSS Statistics to run the analysis and produce the results immediately in
the Viewer window as presented in Output 11.3.
853
ONE-WAY ANOVA is done as a follow-up to see how the results of the
analysis could have been without controlling for the IQ Pre-test
(covariate). For the purpose of having an idea of the effect of Lumosity
training schedule on Intelligence Quotient (IQ) when the influence of the
pre-test has not been eliminated with ANCOVA, execute One-Way
ANOVA for the data. Activate the Data Editor containing the dataset
(Chapter 11 Table 2 Data.sav) exactly as in Figure 11.31.
Click Analyze Compare Means One-Way ANOVA as
shown in Figure 11.46.
854
Figure 11.47 One-Way ANOVA dialog box
Click Post Hoc pushbutton for the One-Way ANOVA: Post Hoc
Multiple Comparisons dialog box to appear as shown in Figure 11.49 in
which you select one of the listed possible tests, Bonferroni checkbox in
this example, and click Continue.
855
Figure 11.49 One-Way ANOVA: Post Hoc Multiple Comparisons dialog box
Figure 11.50 One-Way ANOVA: Options with Descriptive, Homogeneity of variance test,
and Mean plot checked
856
Click Continue to get back to the One-Way ANOVA main menu.
Finally, click OK to have SPSS run the analysis and produce the results as
presented in Section B of Output 11.3.
Please, note that the One-Way ANOVA that is done in the example is not
necessary as it is not part and parcel of ANCOVA. It was only put here to
illustrate in this situation that ANCOVA is the most appropriate statistical
test to apply. Never ever run from performing ANCOVA by applying One-
Way ANOVA in an investigation that requires elimination of specific
covariates. Perform ANCOVA whenever ANCOVA is the test to be used as
I have explained at the beginning of this chapter. In a situation that after
performing the ANCOVA, you executed ANOVA as follow-up test, it is the
ANCOVA output that must be reported for data-based decisions to be made.
857
858
859
860
861
862
863
864
INTERPRETATION OF OUTPUT 11.3 – EFFECT OF
LUMOSITY TRAINING SCHEDULE ON IQ
The Output 11.3 has 13 parts (8 for the ANCOVA proper and 5 for the
Section B [ANOVA follow-up test]) on the effect of Lumosity training
schedule on Intelligence Quotient. They are the:
Univariate Between-Subjects Factors
Descriptive Statistics
Levene’s Test of Equality of Error Variances
Tests of Between-Subjects Effects
Estimated Marginal Means for LumosityTS
Pairwise Comparisons
Univariate Tests
Profile Plots
Oneway Descriptives
Test of Homogeneity of Variances
ANOVA
Post Hoc Tests of Multiple Comparisons with Bonferroni
Means Plots.
Each of these segments of the Output 11.3 is very aptly interpreted.
865
Univariate Analysis of Variance Between-Subjects
Factors
This table merely lists the Value Label (00minpd, 06minpd, 12minpd,
18minpd and 24minpd) respectively for the five levels of LumosityTS
(1.00, 2.00, 3.00, 4.00 and 5.00). The N (number of cases) for each of the
five levels is 30.
Descriptive Statistics
The Descriptive Statistics table has presented the Intelligence Quotient
Post-test (IQPost-test) Mean, Std. Deviation and N for each of the five
levels of LumosityTS. The table could serve as part of the answers to the
research questions. There is equal N of 30 for each level of the independent
variable. The Intelligence Quotient Post-test has a Mean and Std. Deviation
respectively of 107.2333 and 11.02875 for 00minpd, 106.4000 and
11.30334 for 06minpd, 109.6000 and 11.10328 for 12minpd, 106.4333 and
8.79531 for 18minpd, and 109.4333 and 11.20247 for 24minpd. The total
has 150 N, 107.8200 Mean and 10.67901 Std. Deviation.
866
is summarised in this ANCOVA table is whether the five levels or treatment
conditions of Lumosity training schedule (LumosityTS) produce
significantly different scores on the dependent variable (Intelligence
Quotient at the post-test) when the pre-test IQ is controlled for. Information
in the row in the table for the independent variable (Lumosity) takes
adequate care of the main concern of the investigation. The Lumosity row
has shown 164.025 Type III Sum of Squares, 4 df, 41.006 Mean Square,
3.177 F, .016 Sig., and .081 Partial Eta Squared (ɳp2). The .016 Sig. is less
than the classically chosen alpha level of .05. Therefore, the omnibus null
hypothesis which is the first null hypothesis of the experimental study that
“there is no significant effect of Lumosity training schedule on Intelligence
Quotient when the Pre-treatment IQ test (IQPre-test) influence has been
controlled for (held constant)” is rejected. Thus, the alternate hypothesis
that “there is a significant effect of Lumosity training schedule on the
subjects’ Intelligence Quotient (IQ) when the pre-treatment test (Pre-test)
influence has been controlled for (held constant)” is sustained. In other
words, there is a statistically significant evidence that the five LumosityTS
groups (00minpd, 06minpd, 12minpd, 18minpd, and 24minpd) significantly
differ in their means and standard deviations on Intelligence Quotient after
controlling for the effect of the IQ pre-test. The Pairwise Comparisons that
will be discussed in a short while from now will reveal the specific pair or
pairs of Lumosity training schedule that have statistically significant
difference.
867
Partial Eta squared is SSLumosity (164.025); the denominator of the Partial
Eta squared is SSLumosity plus SSError (i.e., 164.025 + 1858.856) that is equal
to 2022.881. Therefore, the Partial Eta Squared is simply 164.025 ÷
2022.881 that is equal to .081 as shown in the Lumosity row of the last
column of the Tests of Between-Subjects Effects table.
Perhaps, you may comprehend the Effect Size in ANCOVA that is
measured with Partial Eta Squared (ɳp2) better in equation form as follows.
Since the Sum of Squares Effect in this case is actually the Sum of
Squares Lumosity, the Equation 11:4 can accordingly be specifically
rewritten as given in Equation 11:5.
The Sum of Squares Lumosity is 164.025; and the Sum of Squares Error
is 1858.856. Therefore, replacing the symbols in Equation 11:5 with their
actual values for the effect of Lumosity training schedule on Post-test IQ
when the Pre-test IQ is controlled for or eliminated, the Partial Eta Squared,
ɳp2, is as in Equation 11:5.
868
between the dependent variable, Intelligence Quotient post-treatment test
scores (IQPost-test) and the covariate, Intelligence Quotient pre-treatment
test scores (IQPre-test) when the effect of the independent variable
Lumosity training schedule (LumosityTS) is statistically controlled for. The
Sig. for the covariate is .000 that is smaller than the .05 alpha; therefore,
association between the dependent variable and the covariate is statistically
significant when Lumosity training schedule is held under control.
In all, the Tests of Between-Subjects Effects table has presented the
Type III Sum of Squares, df, Mean Square, F, Sig., and Partial Eta Squared
(each of these serving as a column) for each of the sources of variation,
namely: Corrected Model, Intercept, IQPre, and Lumosity (each of these
sources serves as a row). The F for each of these sources is significant as
the Sig. for it is lower than the chosen .05 classical level of significance.
The table has equally displayed the Type III Sum of Squares (SSTIII), df,
and Mean Square for Error to be 1858.856, 144 and 212.909, respectively.
While the Total has 1760765.000 SSTIII and 150 df, the Corrected Total
has 16992.140 SSTIII and 149 df.
869
This Pairwise Comparisons table has presented the Post-test Intelligence
Quotient (Post-test IQ) Mean Difference (I-J), Std. Error, Sig., and 95%
Confidence Interval for Difference Lower Bound and Upper Bound. The
Pairwise Comparisons of Post-test IQ is crucially important in the
ANCOVA because each of the other null hypotheses beside the omnibus
one is tested with information in this table. That is, the null hypotheses 2, 3,
4, 5, 6, 7, 8, 9, 10, and 11 are tested independently with information in the
appropriate rows of the Pairwise Comparisons of the Post-test IQ table.
On the whole, every Pairwise Comparison Mean Difference that is
statistically significant, has an asterisk attached to it, with Sig. that is less
than .05, and both the Lower Bound and Upper Bound 95% Confidence
Interval for Difference completely fall either below zero or completely
above zero. A Pairwise Mean Difference that is not significant statistically,
has no asterisk attached, the Sig. is greater than .05, and the Lower Bound
and Upper Bound 95% Confidence Interval for Difference has one side
below zero and the other side of it is above zero. Whenever the Post-test IQ
Mean of (I) LumosityTS is smaller than the Post-test IQ Mean of the (J)
LumosityTS, the Mean Difference has a negative sign as a prefix because
the Mean (J) is subtracted from the Mean (I).
The null hypothesis 2 that “there is no significant difference between
the Intelligence Quotient of those who did not receive Lumosity training
any day (00minpd) and those who received 6 minutes Lumosity training per
day (06minpd) when the Pre-test is controlled for” is retained (not rejected)
because the -1.081 Mean Difference has Sig. of 1.000 that is more than .05
chosen alpha. There is indeed, no significant difference between the IQ of
those who did not receive Lumosity training any day (00minpd) and those
who received 6 minutes Lumosity training per day (06minpd) when the Pre-
test is controlled for.
The null hypothesis 3 that “there is no significant difference between
the Intelligence Quotient of those who did not receive Lumosity training
any day (00minpd) and those who received 12 minutes Lumosity training
per day (12minpd) when the Pre-test is controlled for” is retained (not
rejected) because the Mean Difference of -2.558 has .066 Sig. that is higher
than .05 classically chosen alpha. Therefore, there is indeed, no significant
difference between the Intelligence Quotient of those who did not receive
Lumosity training any day (00minpd) and those who received 12 minutes
Lumosity training per day (12minpd) when the Pre-test is controlled for.
870
The null hypothesis 4 that “there is no significant difference between
the Intelligence Quotient of those who did not receive Lumosity training
any day (00minpd) and those who received 18 minutes Lumosity training
per day (18minpd) when the Pre-test is controlled for” is retained (failed to
reject). Reason for the retention is that the Mean Difference of -1.433 has
1.000 Sig. that is greater than the chosen alpha of .05. Therefore, there is no
significant difference between the Intelligence Quotient of those who did
not receive Lumosity training any day (00minpd) and those who received
18 minutes Lumosity training per day (18minpd) when the Pre-test is
controlled for.
Null hypothesis 5 that “there is no significant difference between the
Intelligence Quotient of those who did not receive Lumosity training any
day (00minpd) and those who received 24 minutes Lumosity training per
day (24minpd) when the Pre-test is controlled for” is retained. Reason for
the retention is that the Mean Difference of .384 has Sig. of 1.000 that is
greater than the .05 chosen alpha. Therefore, there is indeed, no significant
difference between the Intelligence Quotient of those who did not receive
Lumosity training any day (00minpd) and their counterparts who received
24 minutes Lumosity training per day (24minpd) when the Pre-test is
controlled for.
Null hypothesis 6 that “there is no significant difference between the
Intelligence Quotient of those who received 6 minutes Lumosity training
per day (06minpd) and those who received 12 minutes Lumosity training
per day (12minpd) when the Pre-test is controlled for” is retained. The
retention is because the Mean Difference of -1.477 has 1.000 Sig. that is
greater than .05 alpha. There is therefore no significant difference
statistically between the Intelligence Quotient of those who received 6
minutes Lumosity training per day (06minpd) and those who received 12
minutes Lumosity training per day (12minpd) when the Pre-test is
controlled for.
The null hypothesis 7 of “no significant difference between the
Intelligence Quotient of those who received 6 minutes Lumosity training
per day (06minpd) and those who received 18 minutes Lumosity training
per day (18minpd) when the Pre-test is controlled for” is retained. Reason
for retaining the null hypothesis 7 is that the Mean Difference of -.352 has
Sig. of 1.000 that is greater than the classically chosen alpha of .05. There is
therefore, no significant difference between the Intelligence Quotient of
871
those who received 6 minutes Lumosity training per day (06minpd) and
those who received 18 minutes Lumosity training per day (18minpd) when
the Pre-test IQ is controlled for.
The null hypothesis 8 that “no significant difference between the
Intelligence Quotient of those who received 6 minutes Lumosity training
per day (06minpd) and those who received 24 minutes Lumosity training
per day (24minpd) when the Pre-test is controlled for” is retained (failed to
reject). This non-rejection is because the Mean Difference of 1.465 has
1.000 Sig. that is higher than .05 classically chosen alpha. Therefore, there
is, indeed, no significant difference between the Intelligence Quotient of
those who received 6 minutes Lumosity training per day (06minpd) and
those who received 24 minutes Lumosity training per day (24minpd) when
the Pre-test is controlled for.
Null hypothesis 9 of “no significant difference between the Intelligence
Quotient of those who received 12 minutes Lumosity training per day
(12minpd) and those who received 18 minutes Lumosity training per day
(18minpd) when the Pre-test is controlled for” is retained (failed to reject).
The reason for retaining the null hypothesis is that the Mean Difference of
1.125 has Sig. of 1.000 which is greater than the chosen alpha of .05.
Therefore, there is indeed, no statistically significant difference between the
Intelligence Quotient of those who received 12 minutes Lumosity training
per day (12minpd) and those who received 18 minutes Lumosity training
per day (18minpd) when the Pre-test IQ is eliminated or held constant.
The null hypothesis 10 that “there is no significant difference between
the Intelligence Quotient of those who received 12 minutes Lumosity
training per day (12minpd) and those who received 24 minutes Lumosity
training per day (24minpd) when the Pre-test is controlled for” is rejected.
This rejection is because the Mean Difference of 2.942* has Sig. of .019
which is less than the classically chosen alpha of .05. Therefore, there is a
statistically significant difference between the Intelligence Quotient of those
who received 12 minutes Lumosity training per day (12minpd) and those
who received 24 minutes Lumosity training per day (24minpd) when the
Pre-test is controlled for. The significant mean difference is in favour of
those who had 12 minutes Lumosity training per day. You can see that the
mean difference has asterisk attached to it as a sign that it is significant at
the .05 alpha level. Furthermore, the 95% Confidence Interval for
Difference Lower Bound and Upper Bound are both above zero, indicating
872
significance of the mean difference. It can quickly be noted that of all the
pairwise mean difference comparisons, this is the only one that is
significant statistically.
The null hypothesis 11 that “there is no significant difference between
the Intelligence Quotient of those who received 18 minutes Lumosity
training per day (18minpd) and those who received 24 minutes Lumosity
training per day (24minpd) when the Pre-test is controlled for” is retained
(failed to reject). Reason for the failure to reject null hypothesis 11 is that
the Mean Difference of 1.818 has Sig. of .547 that is greater than the
classically chosen alpha of .05. Thus, the alternate hypothesis that “there is
a significant difference between the Intelligence Quotient of those who
received 18 minutes Lumosity training per day (18minpd) and those who
received 24 minutes Lumosity training per day (24minpd) when the Pre-test
is controlled for” is discarded. What seems to be like a mean difference
between the two levels of Lumosity training (18 minutes per day and 24
minutes per day) with regard to the post-test Intelligence Quotient when the
pre-test is controlled for, is merely a function of chance that is not
consistent; and cannot qualify for a significant difference.
In addition to testing 10 of the 11 null hypotheses, information in the
Pairwise Comparisons of Post-test IQ can very suitably be used to answer
the 11 research questions. The Pairwise Mean Difference in each
comparison, is an answer to the corresponding research question. For
research questions 2 and 3 for instance, the Mean Difference of -1.081 and
-2.558, respectively, could serve as the answers as illustrated in Table 11.2a.
873
when the influence of the IQ pre-test has M=109.440
been held constant for?) SD=11.10328
18minpd
M=108.316
SD=11.20247
24minpd
M=106.498
SD=10.67901
ii. What is the difference between the IQ of -1.081 Not
those who did not receive Lumosity significant
training any day (00minpd) and those p > .05
who received 6 minutes Lumosity
training per day (06minpd) when the
Pre-test is controlled for?
iii. What is the difference between the IQ of -2.558 Not
those who did not receive Lumosity significant
training any day (00minpd) and those p > .05
who received 12 minutes Lumosity
training per day (12minpd) when the
Pre-test is controlled for?
iv. What is the difference between the IQ of -1.433 Not
those who did not receive Lumosity significant
training any day (00minpd) and those p > .05
who received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is controlled for?
v. What is the difference between the IQ of .384 Not
those who did not receive Lumosity significant
training any day (00minpd) and those p > .05
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is controlled for?
vi. What is the difference between the IQ of -1.477 Not
those who received 6 minutes Lumosity significant
training per day (06minpd) and those p > .05
874
who received 12 minutes Lumosity
training per day (12minpd) when the
Pre-test is controlled for?
vii. What is the difference between the IQ of -.352 Not
those who received 6 minutes Lumosity significant
training per day (06minpd) and those p > .05
who received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is controlled for?
viii. What is the difference between the IQ of 1.465 Not
those who received 6 minutes Lumosity significant
training per day (06minpd) and those p > .05
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is controlled for?
ix. What is the difference between the IQ of 1.125 Not
those who received 12 minutes Lumosity significant
training per day (12minpd) and those p > .05
who received 18 minutes Lumosity
training per day (18minpd) when the
Pre-test is controlled for?
x. What is the difference between the IQ of 2.942 Significant
those who received 12 minutes Lumosity p < .05
training per day (12minpd) and those
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is controlled for?
xi. What is the difference between the IQ of 1.818 Not
those who received 18 minutes Lumosity significant
training per day (18minpd) and those p > .05
who received 24 minutes Lumosity
training per day (24minpd) when the
Pre-test is controlled for?
Univariate Tests
875
The Univariate Tests table presents the Post Test IQ Sum of Squares, df,
Mean Square, F, Sig., and Partial Eta Squared for Contrast and Error. The F
is 3.177 with .016 Sig. The F of 3.177 is statistically significant as the Sig.
of .016 is less than .05 alpha; and the Partial Eta Squared is .081. That is,
the omnibus null hypothesis tested by the Univariate Tests should be
rejected [F(4, 144) = 3.177, p < .05, ɳp2 = .081]. This F indicates the effect
of LumosityTS on the basis of linearly independent pairwise comparisons
among the estimated marginal means.
Profile Plots
The Profile Plots serve as pictorial illustration of answers to the research
questions. The Plots indicate the Estimated Marginal Means of Post-test
Intelligence Quotient as a function of the Lumosity training schedule of
00minpd, 06minpd, 12minpd, 18minpd, and 24minpd. The covariates
appearing in the model on which the Profile Plots is based was evaluated at
the following values: Pre-test IQ = 105.9667.
876
from 106.4000 to 109.6000 in no particular order. While the mean for Total
during Pre-test IQ is 105.9667, the mean for Total during Post-test IQ is
107.8200. All the other values in the table can easily be read and interpreted
in a similar manner.
ANOVA
The ANOVA table has shown that for Pre-test IQ, the Sum of Squares
Between Groups is 483.067, df Between Groups is 4, and the Mean Square
Between Groups is 120.767. The Within Groups at Pre-test IQ has
16189.767 Sum of Squares, 145 df, and 111.654 Mean Square. The Total
Sum of Squares and df are respectively 16672.833 and 149. The F is 1.082
with .368 Sig. Since the Sig. of .368 is greater than the classically chosen
.05 alpha, the omnibus null hypothesis of no significant effect of Lumosity
training schedule on Intelligence Quotient at the Pre-test is retained (not
rejected). The Effect Size, measured with Partial Eta Squared is .029 that is
got by dividing the Sum of Squares Between Groups (SSB) by the Total
Sum of Squares (SST). That is, Partial Eta Squared (ɳp2) for ANOVA in
equation as rationalized earlier in this chapter (Equation 11:3) is:
That is, Lumosity training schedule accounts for just 2.9% of the
variance in Pre-test Intelligence Quotient (Pre-test IQ). With the complete
randomization done to constitute the five groups to arrive at the treatment
condition that each of the groups was to have, the IQ of the subjects across
the groups before exposure to the experimental treatment was bound to be
877
about equal, not different statistically exactly as confirmed by this result
that the p > .05.
At the Post-test IQ, the Between Groups has 301.640 Sum of Squares, 4
df, and 75.410 Mean Square. The Within Groups at the Post-test IQ has
16690.500 Sum of Squares, 145 df, and 115.107 Mean Square. The Total
Sum of Squares and df are 16992.140 and 149, respectively. At the Post-test
IQ, the F is .655 with .624 Sig. The Sig. (.624) is greater than the
classically chosen alpha of .05. Therefore, the omnibus null hypothesis of
no significant effect of Lumosity training schedule on Intelligence Quotient
at the Post-test is retained (not rejected). The Effect Size as measured with
Partial Eta Squared is .018. In equation form, Partial Eta Squared (ɳp2) for
One-Way ANOVA is as given earlier as Equation 11:3:
The Effect Size (Partial Eta Squared) implies that Lumosity training
schedule accounts for only 1.8% of variance in the Post-test Intelligence
Quotient (Post-test IQ). It could be deduced at this point as a confirmatory
evidence that if not for the very special role that ANCOVA plays, there
could not have been a significant effect of Lumosity training schedule on
Intelligence Quotient at the Post-test when the influence of the Pre-test has
been controlled for as revealed by the ANCOVA results where the F
(3.177) was significant (p < .05). I mean that while with the ANCOVA, F
(3.177) was significant (p < .05); with ANOVA, F (.655) is insignificant (p
> .05) because ANCOVA statistically increases the power and precision of
the F test by increasing the F ratio via eradication of the effect of the
concomitant variable that is called covariate from the dependent variable.
878
significant as there is no Mean Difference with asterisk because for each
Mean Difference, the Sig. is greater than .05 alpha.
At the Post-test IQ too, none of the Pairwise Comparisons is significant
statistically because the Sig. in each of the Pairwise Comparisons is greater
than the classically chosen alpha of .05. Consequently, there is no asterisk
attached to any of the Pairwise Mean Differences. The total absence of
asterisk in the Mean Differences show that no single mean difference is
significant at the .05 alpha level. That is, Lumosity training schedule has no
significant effect on Intelligence Quotient, both at the Pre-test IQ and at the
Post-test IQ.
Note that while the Multiple Comparisons results of the Post-test
Intelligence Quotient in the ANOVA was not precise and powerful enough
to depict a significant Mean Difference between any pair of the means; the
Pairwise Comparisons in the ANCOVA was powerful and precise enough to
spot a statistically significant difference among the Post-test means when
the Pre-test IQ was held constant, controlled for or eliminated. The mean
difference between 12minpd and 24minpd of Lumosity training schedules is
significant in the ANCOVA Post-test IQ, but it is not significant in the
ANOVA Post-test IQ.
Means Plots
There are two Means Plots, one for Pre-test IQ and the other for Post-test
IQ. The first Means Plots is for the Pre-test IQ and it shows that for each of
the five levels of LumosityTS, the mean is approximately within 104 and
109. The second Means Plots clearly shows that the means range closely
from approximately 106.40 for 06minpd Lumosity training group to 109.60
for 12minpd Lumosity training group. The IQ means do not vary widely
enough for any significant difference. So, conclusively, Lumosity training
schedule has no significant effect on the Intelligence Quotient. The pattern
of the curves in the two Means Plots show similarity which depicts
association or relationship between the Pre-test Intelligence Quotient (Pre-
test IQ) and the Post-test Intelligence Quotient (Post-test IQ) across the five
experimental conditions.
879
ANCOVA was executed for the effect of Lumosity training schedule
(LumosityTS) on Intelligence Quotient Post-test (Post-test IQ) when the
effect of the Intelligence Quotient Pre-test (Pre-test IQ) is controlled for,
held constant or eliminated. The assumptions for execution of ANCOVA
were met as the Levene’s Test of Equality of Error Variances was not
significant (p > .05) and the covariate (Pre-test IQ) had significant linear
relationship with the dependent variable (Post-test IQ) [F(1, 144) =
1148.963, p < .05, ɳp2 = .889]. The ANCOVA of the main concern in the
study showed a statistically significant effect of LumosityTS on Intelligence
Quotient when the Pre-test IQ is controlled for [F(4, 144) = 3.177, p < .05,
ɳp2 = .081]. To ascertain the pairwise mean differences that were significant
statistically, Pairwise Comparisons done revealed significant Mean
Difference (MD) for only one pair, 12minpd and 24minpd, [MD = 2.942, p
< .05]. All the other Pairwise Comparisons showed total lack of significant
Mean Difference (MD). For instance, between 00minpd and 06minpd [MD
= -1.081, p > .05]; 00minpd and 12minpd [MD = -2.558, p > .05]; and
00minpd and 18minpd [MD = -1.433, p > .05]; 00minpd and 24minpd [MD
= .384, p > .05]; between 06minpd and 12minpd [MD = -1.477, p > .05];
06minpd and 18minpd [MD = -.352, p > .05]; 06minpd and 24minpd [MD
= 1.465, p > .05]; 12minpd and 18minpd [MD = 1.125, p > .05]; and
between 18minpd and 24minpd [MD = 1.818, p > .05]. Profile Plots was
drawn to pictorially illustrate the Intelligence Quotient clustered Mean
Differences across the five levels of LumosityTS.
The One-Way ANOVA that was done was as a follow-up to ascertain the
results supposing the Pre-test IQ was not controlled for. Results showed that
effect of LumosityTS on the Pre-test IQ is not significant at all [F(4, 145) =
1.082, p > .05, ɳp2 = .029]. Post Hoc Multiple Comparisons confirmed that
all of the pairwise comparisons were not significant as each Sig. was greater
than .05 alpha. Means Plots for the Intelligence Quotient at the Pre-test
showed means that ranged from approximately 104.00 to approximately
109.03.
Similarly, the effect of LumosityTS on the Post-test IQ was not
significant statistically [F(4, 145) = .655, p > .05, ɳp2 = .018]. Post Hoc
Multiple Comparisons showed total lack of significant Pairwise Mean
Difference (MD) between 00minpd and 06minpd [MD = -.83333, p > .05];
00minpd and 12minpd [MD = -2.36667, p > .05]; and 00minpd and
880
18minpd [MD = .80000, p > .05]; 00minpd and 24minpd [MD = -2.20000,
p > .05]. There was also lack of statistically significant Mean Difference
(MD) between 06minpd and 12minpd [MD = -3.20000, p > .05]; 06minpd
and 18minpd [MD = -.03333, p > .05]; 06minpd and 24minpd [MD =
-3.03333, p > .05]; 12minpd and 18minpd [MD = 3.16667, p > .05];
12minpd and 24minpd [MD = .16667, p > .05]. The Mean Difference (MD)
between 18minpd and 24minpd too was not significant [MD = -3.00000, p
> .05]. Means Plots was drawn to pictorially illustrate the Intelligence
Quotient Means across the five levels of LumosityTS. It showed that the
means are close by, ranging only from 106.40 for 06minpd to 109.60 for
12minpd without consistency of direction across the five LumosityTS
levels. The curves in the Means Plots for Pre-test IQ and Post-test IQ were
alike to indicate high positive correlation or association between the Pre-
test IQ and Post-test IQ.
Step 6. View the output once more, taking notes of necessary details,
particularly with regard to answering the research questions and testing the
null hypotheses.
Step 7. Save the output. Click Save this document icon in the Toolbar.
It produces the Save Output As dialog box. Type the suitable name for the
document, Chapter 11 Output 3.spv in this case, in the File name box and
click Save. With this, the output is well saved for future reference.
Step 8. Print the output and the data. The printing procedure is very simple
as explained repeatedly in previous chapters. Perform similar actions of
clicking the Print tool in the Toolbar; and clicking OK in the Print dialog
box.
Step 9. Exit IBM SPSS Statistics simply by closing each of the SPSS
windows that are open on your system.
Step 10. Shut down the computer. Click Start Power Shut
down.
881
ANCOVA WITH IBM SPSS SYNTAX FOR
LUMOSITYTS ON IQPOST
The SPSS Programming Language, Commands Syntax, for performing
Analysis of Covariance (ANCOVA) is given in Syntax 11.2. All that is
required of you is to unmistakably type the Syntax with the aid of SPSS
prompting into the IBM SPSS Statistics Syntax Editor. Whenever you are
to execute ANCOVA for data on some other variables, only the names and
labels of the variables that you will have to substitute with the variable
names in this example. Every other thing remains the same as shown in
Syntax 11.2. The Syntax 11.2 contains the exact SPSS Commands that can
be used to execute the ANCOVA in this example that deals with the effect
of Lumosity Training Schedule on Intelligence Quotient (IQ).
882
The first part of the Syntax, from the beginning to the first period (full
stop), is used to execute the ANCOVA that is the core concern of this
example (effect of Lumosity Training on Intelligence Quotient when the
influence of the pre-treatment IQ test is completely eliminated, held
constant, or controlled for by statistically treating the IQ Pre-test as
covariate). The second part of the Syntax is used to perform One-Way
ANOVA (which is not the concern of this example really) but given here
merely to illustrate what the output could have been if not for the
elimination of the pre-test effect with the special statistical test, called
ANCOVA.
Switch on the computer and allow it to fully boot. Launch IBM SPSS
Statistics by clicking its icon on the computer. Wait until it gets
completely set and close the superimposed dialog box on the Data Editor.
Key in the data carefully without any mistake; after which you name and
label the variables accordingly. But since the data in question have already
been entered and saved earlier, simply retrieve the data. The data retrieval is
done by selecting File Open Data for the Open Data dialog
box to appear. Alternatively, it can be done by clicking the Open data
document icon in the Toolbar for the Open Data dialog box to appear.
Look for the name with which the data document was saved, Chapter 11
Table 2 Data.sav, click it and press the Enter key in the Keyboard. This
opens the data file as earlier exhibited in Figure 11.31.
Move cursor to and click Window to have its pulldown menu. In the
Window pulldown menu, click Go to Designated Syntax Window (see
Figure 11.51).
883
By clicking the Go to Designated Syntax Window, the IBM SPSS
Statistics Syntax Editor gets activated as shown in Figure 11.52.
Alternatively, click File, select New, and click Syntax to have the IBM
SPSS Statistics Syntax Editor. It is in this window that the Syntax 11.2,
given earlier, must be typed in accurately to perform the analysis.
Key in the Syntax 11.2 with all accuracy. It is very simple to accurately
enter the Syntax because, once you type the first alphabet of each word or
phrase, SPSS automatically prompts you by providing words or phrases in
the Syntax Command that begin with the alphabet. You simply click the
very word or phased that you wanted to type. It is only the names and labels
of your variables that require you to type as you have done in the Variable
View. When you correctly key in the entire Syntax, the IBM SPSS
Statistics Syntax Editor will look like Figure 11.53.
884
Figure 11.53 IBM SPSS Statistics Syntax Editor with the entered syntax
To have SPSS perform the analysis and produce the output in the
Viewer window, click Run and in its pulldown menu, click All. With this,
IBM SPSS Statistics performs the whole analysis and produces the results
in the IBM SPSS Statistics Viewer window as presented in Output 11.4.
885
886
887
888
889
890
891
892
INTERPRETATION OF OUTPUT 11.4
Output 11.4 is the same as Output 11.3. While the latter (Output 11.3) was
arrived at through IBM SPSS Statistics dialog boxes point and click
method, the former (Output 11.4) was a product of IBM SPSS Statistics
Syntax application on the data. Therefore, the same interpretation for
Output 11.3 suitably goes for the Output 11.4. Any hypothesis that is
rejected, using Output 11.3 is also rejected with the use of Output 11.4. The
only Pairwise Multiple Comparison that is significant in Output 11.3 is the
one that is statistically significant in the Output 11.4; while the Pairwise
Multiple Comparisons that are not significant in Output 11.3 are also not
significant in Output 11.4.
Follow the remaining SPSS data analysis protocols to eventually shut
down your system.
View output
Save the output as Chapter 11 Output 4.spv and the Syntax as Chapter
11 Syntax 2.sps
893
Print the Output and Syntax
894
Appendix A: Minimum suitable sample
size from population
895
190.00 400.00
200.00 450.00
210.00 500.00
220.00 500.00
130.00 600.00
140.00 650.00
150.00 700.00
260.00 800.00
270.00 900.00
280.00 980.00
290.00 1,000.00
300.00 1,200.00
310.00 1,400.00
320.00 1,600.00
330.00 1,900.00
340.00 2,300.00
350.00 3,000.00
360.00 4,000.00
370.00 5,000.00
380.00 10,000.00
390.00 1,000,000.00
400.00 5,000,000.00
410.00 10,000,000.00
420.00 100,000,000.00
420.00 1,000,000,000.00
420.00 100,000,000,000.00
896
420.00 1,000,000,000,000.00
897
Appendix B: IBM SPSS SYNTAX
Summary
898
MEANS TABLES=ATEB BY MaritalStatus
/CELLS=MEAN COUNT STDDEV MEDIAN GMEDIAN SEMEAN
SUM MIN
MAX RANGE FIRST LAST VAR KURT SEKURT SKEW
SESKEW HARMONIC GEOMETRIC SPCT NPCT.
899
/CRITERIA=CI(.95).
900
Syntax 6.2 Chapter 6 Syntax 2.sps
UNIANOVA LumEfficacy BY BulbTypes
/METHOD=SSTYPE(3)
/INTERCEPT=INCLUDE
/POSTHOC=BulbTypes(SCHEFFE
BONFERRONI QREGW)
/PLOT=PROFILE(BulbTypes) TYPE=BAR
ERRORBAR=NO MEANREFERENCE=NO
/EMMEANS=TABLES(BulbTypes)
COMPARE ADJ(LSD)
/PRINT ETASQ DESCRIPTIVE
/CRITERIA=ALPHA(.05)
/DESIGN=BulbTypes.
901
/EMMEANS=TABLES(Interventions) COMPARE
ADJ(BONFERRONI)
/PRINT=DESCRIPTIVE ETASQ
/CRITERIA=ALPHA(.05)
/WSDESIGN=Interventions.
T-TEST PAIRS=Baseline Baseline Baseline Physiological
Physiological Social WITH Physiological
Social Psychological Social Psychological
Psychological (PAIRED)
/CRITERIA=CI(.9500)
/MISSING=ANALYSIS.
902
/MEASURE=LTMemory
/METHOD=SSTYPE(3)
/PLOT=PROFILE(MemDrug*LearnStrategy) TYPE=LINE
ERRORBAR=NO MEANREFERENCE=NO YAXIS=AUTO
/EMMEANS=TABLES(MemDrug) COMPARE ADJ(BONFERRONI)
/EMMEANS=TABLES(LearnStrategy) COMPARE
ADJ(BONFERRONI)
/EMMEANS=TABLES(MemDrug*LearnStrategy)
/PRINT=DESCRIPTIVE ETASQ
/CRITERIA=ALPHA(.05)
/WSDESIGN=MemDrug LearnStrategy MemDrug*LearnStrategy.
903
/EMMEANS=TABLES(Stem) COMPARE
ADJ(BONFERRONI)
/EMMEANS=TABLES(Discipline*Stem)
/PRINT=DESCRIPTIVE ETASQ
/CRITERIA=ALPHA(.05)
/WSDESIGN=Stem
/DESIGN=Discipline.
MANOVA Science Technology Engineering Maths BY
Discipline(1,3)
/WSFACTORS=Stem(4)
/WSDESIGN=MWITHIN Stem(1) MWITHIN Stem(2)
MWITHIN Stem(3) MWITHIN Stem(4).
MANOVA Science Technology Engineering Maths BY
Discipline(1,3)
/WSFACTORS=Stem(4)
/DESIGN=MWITHIN Discipline(1) MWITHIN Discipline(2)
MWITHIN Discipline(3).
904
UNIANOVA IQPost BY Lumosity WITH IQPre
/METHOD=SSTYPE(3)
/INTERCEPT=INCLUDE
/PLOT=PROFILE(Lumosity) TYPE=LINE ERRORBAR=NO
MEANREFERENCE=NO YAXIS=AUTO
/EMMEANS=TABLES(Lumosity) WITH(IQPre=MEAN)
COMPARE ADJ(BONFERRONI)
/PRINT ETASQ DESCRIPTIVE HOMOGENEITY
/CRITERIA=ALPHA(.05)
/DESIGN=IQPre Lumosity.
ONEWAY IQPre IQPost BY Lumosity
/STATISTICS DESCRIPTIVES HOMOGENEITY
/PLOT MEANS
/MISSING ANALYSIS
/POSTHOC=BONFERRONI ALPHA(0.05).
905
REFERENCES
906
Herrnstein, R. J. & Murray, C. (1994). The bell curve intelligence and class
structure in American life. New York: THE FREE PRESS.
Ho, R. (2014). Handbook of univariate and multivariate data analysis with
IBM SPSS (2nd ed.). London: CRC Press – Taylor & Francis Group.
Howell, D. C. (2002). Statistical Methods for Psychology. Australia:
Thomson Wadsworth.
Howell, D. C. (2006). Statistical methods for psychology. Australia:
Thomson Wadsworth Publishing.
Howell, D. S. (2013) Statistical methods for psychology. Belmont, CA:
Wadsworth, Centage Learning.
Howell, D. S. (2016). Fundamental statistics for the behavioural sciences
(9th ed.). Belmont, CA: Wadsworth, Centage.
Hunt, E. (2011). Human intelligence. New York: Cambridge University
Press.
IBM (2019). 2019 IBM Annual Report. https://www.ibm.com/annualreport
IBM (2020). IBM SPSS Statistics. https://www.ibm.com/products/spss-
statistics
IBM Corp. (2019). IBM SPSS Statistics 26 Command Syntax Reference.
USA: IBM Corporation.
IBM Corp. (2019). IBM SPSS Statistics Grad Pack 26.0 PREMIUM. USA:
IBM Corporation.
Johnson, R. A. & Wichern, D. W. (2014). Applied multivariate statistical
analysis (6th ed.). Delhi: PHI Learning Private Limited.
Kpolovie, P. J. & Onoshagbegbe, E. S. (2017). Research Productivity: h-
index and i10-index of academics in Nigerian universities. International
Journal of Quantitative and Qualitative Research Methods. 5(2), 62-
123. Retrieved from http://www.eajournals.org/wp-
content/uploads/Research-Productivity-h-Index-and-I10-Index-of-
Academics-in-Nigerian-Universities.pdf
Kpolovie, P. J. (2010). Advanced research methods. Owerri, Imo State:
Springfield Publishers Ltd.
Kpolovie, P. J. (2011). Cognitive enhancement: Effects of Lumosity
Training and Brain-Boosting Food on Learning. Owerri: Springfield
Publishers Ltd.
Kpolovie, P. J. (2011). Statistical techniques for advanced research. Owerri:
Imo State: Springfield Publishers Ltd.
907
Kpolovie, P. J. (2011a). Hypothesis postulation: The what, why, how, and
which? Journal of Education in Developing Areas (JEDA).
https://www.scribd.com/doc/59132454/KPOLOVIE-Hypothesis-
Postulation
Kpolovie, P. J. (2012). Education reforms without evaluation designs:
Nigeria at risk. Owerri, Imo State: Springfield Publishers Ltd.
Kpolovie, P. J. (2014). Test, measurement and evaluation in education.
Owerri: Springfield Publishers.
Kpolovie, P. J. (2015). Indispensability of intelligence testing in the
repositioning and revitalization of Nigerian education. Multidisciplinary
Journal of Research Development. 6(4), 1-11.
http://www.globalacademicgroup.com/journals/nard/Kpolovie.pdf.
Kpolovie, P. J. (2016). Excellent research methods. Indiana, USA: Partridge
Publishing.
Kpolovie, P. J. (2017). Intelligence and academic achievement: A
longitudinal survey. International Journal of Health and Psychology
Research. 5(1), 33-70. http://www.eajournals.org/wp-
content/uploads/Intelligence-and-Academic-Achievement-A-
Longitudinal-Survey.pdf
Kpolovie, P. J. (2017). Statistical analysis with SPSS for research. London:
ECRTD Publication.
Kpolovie, P. J. (2018). Statistical approaches in excellent research methods.
Indiana, USA: Partridge Publishing.
Kpolovie, P. J. (2018a). Fluid Intelligence and Lumosity Training Schedule:
Any effect? International Journal of Humanities, Social Sciences and
Education (IJHSSE). 5(10), 113-144.
https://www.arcjournals.org/pdfs/ijhsse/v5-i10/11.pdf
Kpolovie, P. J. (2018b). Effect of Lumosity Training Schedule on Recall of
Nonsense Syllables. Advances in Social Science Research Journal.
5(10), 206-234.
https://journals.scholarpublishing.org/index.php/ASSRJ/article/view/529
0/3288
Kpolovie, P. J. (2018c). Multiple prediction of Research Productivity: h-
Index. European Journal of Computer Science and Information
Technology. 6(5), 9-36. http://www.eajournals.org/wp-
content/uploads/Multiple-Prediction-of-Research-Productivity-H-
Index.pdf
908
Kpolovie, P. J. and Lale, N. E. S. (2017). Globalization and adaptation of
university curriculum with LMSs in the changing world. European
Journal of Computer Science and Information Technology. 5(2), 28-89.
Retrieved from http://www.eajournals.org/wp-
content/uploads/Globalization-and-Adaptation-of-University-
Curriculum-to-LMSS-with-the-Changing-World.pdf
Kpolovie, P. J. and Ololube N. P. (2020). Effect of types of light bulb on
luminous efficacy.
Kpolovie, P. J., Oshodi, P. O., Iwuchukwu, H. (2016). Continental inequities
in Life Expectancy. European Journal of Biology and Medical Science
Research. 4(6), 30-47. Retrieved from http://www.eajournals.org/wp-
content/uploads/Continental-Inequities-In-Life-Expectancy.pdf
Kpolovie, P. J.; Joe, A. I.; Okoto, T. (2014). Academic achievement
prediction: Role of interest in learning and attitude towards school.
International Journal of Humanities, Social Sciences and Education
(IJHSSE). 1(11), 73-100. https://www.arcjournals.org/pdfs/ijhsse/v1-
i11/10.pdf
Landau, S. & Everitt, B. S. (2004). A handbook of statistical analysis using
SPSS. London: Chapman & Hall/CRC Press LLC.
Mowlaee, N. (2019). Civil Engineering salaries state by state. Interesting
Engineering. https://interestingengineering.com/civil-engineering-
salaries-state-by-state
Ololube, N. P., Kpolovie, P. J. & Makewa, L. N. (2015). Handbook of
research on enhancing teacher education with advanced instructional
technologies. PA, USA: Information Science References, IGI Global.
http://www.igi-global.com/book/handbook-researchenhancing-teacher-
education/120264.
Penn State Center for Nanoscale Sciences (2016). Light bulb efficiency.
Retrieved from https://www.mrsec.psu.edu/content/light-bulb-efficiency
Times Higher Education World University Rankings (2018). Best
Universities in India 2018. Retrieved from
https://www.timeshighereducation.com/student/best-universities/best-
universities-india.
Vogt, W. P. & Johnson, R. B. (2016). The SAGE dictionary of statistics &
methodology – A nontechnical guide for the social sciences (5th ed.).
Loa Angeles: SAGE.
909
Please Leave a Review on Amazon
If this book has enabled you to perform some statistical analysis efficiently,
I’d be very grateful if you took a few minutes to write an honest review on
Amazon.
Your review would be an encouragement for the actualization of the
dream of making more individuals skilful in data analysis with SPSS.
Thank you very much.
910