Professional Documents
Culture Documents
SPSS Reporting With Mediation
SPSS Reporting With Mediation
a. Basic Reporting
Age
Between 20-30 32 21.3
Between 31-40 45 30.0
Between 41-50 37 24.7
50 and above 36 24.0
Qualification
Diploma 11 7.3
Degree 17 11.3
Masters 39 26.0
Doctorate 50 33.3
Others 33 22.0
Sample Explanation
Table 1 shows the descriptive profile of the respondents. A total of 150 respondents
participated in this research. Researchers found that males are the majority of respondents
which consists of 83 (54.7%) while females are 68 (45.3%). While the age group between 31-
40 represents the largest group in this group which is 45 respondents (30%) followed by the
age group between 41-50 which is 37 respondents (24.7%) and the age group 50 and above
which is 36 respondents (24%) ) followed by the age group between 20-30 as many as 32
respondents (21.3%). The age group between 20-30 is the smallest group. In terms of
Qualification, the group with a Doctorate qualification is the largest group found to
participate in this research which is 50 respondents (33.3%) followed by Masters which is 39
respondents (26.0%), first degree 17 respondents (11.3%), other qualifications 33 respondents
(22%) and the diploma of 11 respondents (7.3%). Diploma -educated respondents are the
least group in this research. The ‘others’ category are those with education with a
professional qualification background (5 respondents), vocational (12 respondents) and
certificate level (16 respondents) respectively.
Students’ Notes
Students must explain the basic profile of the respondents and interpret Mean and Standard
Deviation of the nominal data.
Min and Max represent the number of categorical data that you set in the questionnaires.
The mean (average) of a data set is found by adding all numbers in the data set and then
dividing by the number of values in the set.
Standard deviation is a mathematical tool to help you to assess how far the values are
spread above and below the mean. A high standard deviation (more than 1) shows that the
data is widely spread (less reliable) and a low standard deviation (close to zero, not more
than 1) shows that the data are clustered closely around the mean (more reliable).
-1 Min 1
0 Gender Age / Qualification
Standard Deviation (SD) in table 1 explain the average amount of variability in your data set.
It tells you, on average, how far each score lies from the mean. If you look at gender, the data
is skewed to one site but still reliable as it is closed to Zero but not more than 1. Whereas, the
data for age and qualification is less reliable because more than 1. Since it is less reliable, we
will use the non-parametric test to access Age and Qualification only.
Sample Explanation
Table 2 shows the minimum and maximum scores for each continuous variable. Mean
technology acceptance score is the highest which is M = 4.32, SD = .63, followed by
facilitation M = 4.10, SD = .65, perceived confidence M = 3.71, SD = .65 and the lowest is
Cost Benefits M = 2.33 , SD = 88. All standard deviation results show that it is close to the
mean score. This gives an indication that all the variables measured have small distributions.
The mean data for cost benefit indicates that most respondents tend to disagree that cost
benefit is the cause of Technology Acceptance, while the mean facilitation indicates that
facilitation is the main cause of technology acceptance.
A. Normality
Assessing the normality of residuals is important in multiple regression analysis for several
reasons:
1. Assumption of ordinary least squares (OLS): OLS regression assumes that the residuals
(i.e., the differences between the observed values and the predicted values) are normally
distributed. Violation of this assumption can lead to biased parameter estimates and
inaccurate hypothesis testing.
2. Validity of hypothesis tests and confidence intervals: Hypothesis tests and confidence
intervals are based on assumptions of normality. If the residuals are not normally distributed,
the p-values and confidence intervals may be inaccurate, leading to incorrect conclusions.
3. Influential outliers: Non-normality in the residuals may indicate the presence of influential
outliers. These outliers can have a disproportionate impact on estimated regression
coefficients, potentially distorting the results. Identifying and addressing influential outliers is
essential for obtaining reliable estimates.
4. Model fit and predictive accuracy: The assumption of normality is also related to model fit
and predictive accuracy. If the residuals are not normally distributed, it may suggest that the
underlying model does not capture the true data-generating process adequately. Additionally,
non-normal residuals can affect the accuracy of predicted values and may lead to biased
predictions.
To assess normality, you can use diagnostic plots such as a histogram or a Q-Q plot to
visually inspect the distribution of the residuals. Alternatively, statistical tests, such as the
Shapiro-Wilk test or the Kolmogorov-Smirnov test, can be used to formally test for
normality. If the residuals deviate substantially from normality, you may need to consider
transformations of the variables or consider alternative regression methods such as robust
regression that can handle non-normal residuals.
a. Advanced Reporting
Based on the analysis, the data is not significant p>.05 (n=150). Therefore, researcher
conclude that the data is normally distributed.
b. Advanced Reporting
The data in normality test using Shapiro-Wilk and Kolmogorov-Smirnov fulfil the basic
assumption of performing parametric tests after the outliers are detected and removed. Based
on the analysis, the data is not significant p>.05 (n=150). Therefore, researcher conclude that
the data is normally distributed.
Step-by-Step Validity Analysis
To perform a validity analysis in SPSS 16, you can follow these step-by-step instructions:
1. Start SPSS and open your dataset: Launch SPSS 16 and open the dataset that contains the
variables you want to analyze for validity.
2. Determine the variables to analyze: Identify the variables or items you want to assess for
validity. These could be Likert-scale items, continuous variables, or any other measurement
scales.
3. Assess internal consistency using Cronbach's alpha: If you are analyzing a scale or a set of
related items, you can assess internal consistency using Cronbach's alpha. Go to "Analyze" >
"Scale" > "Reliability Analysis". In the "Reliability Analysis" window, select the items you
want to analyze and move them into the "Items" box. Choose the appropriate settings, such as
the desired method for handling missing data, and click "OK". SPSS will generate the
reliability statistics, including Cronbach's alpha, in the output window.
4. Conduct exploratory factor analysis (EFA): If you want to analyze the underlying factor
structure of your items, you can perform exploratory factor analysis. Go to "Analyze" >
"Data Reduction" > "Factor". In the "Factor" window, select the items you want to analyze
and move them into the "Variables" box. Choose the appropriate settings, such as the
extraction method (e.g., principal components analysis) and rotation method (e.g., varimax),
and click "OK". SPSS will generate the factor analysis output, including factor loadings and
eigenvalues.
For details step-by step and result interpretation for KMO and Bartlett’s Test, refer to
Factor Analysis in SPSS - Reporting and Interpreting Results
(onlinespss.com)
5. Assess convergent (p<0.05) and discriminant validity (between 0.3-0.7 or 0.3 and above):
To assess the convergent validity of your scale, examine the factor loadings of each item on
its corresponding factor in the factor analysis output. Generally, factor loadings above 0.30
are considered acceptable. To assess discriminant validity, compare the item loadings across
factors and ensure they are significantly higher on their own factor compared to other factors.
Use this link to assess validity on Instrument (Convergent and divergent validity) using
Bivariate: How to Test Validity questionnaire Using SPSS - SPSS
Tests; SPSS 04: Convergent and Discriminant Validity of Scale
- YouTube; How to analyze discriminant and convergent validity
of Likert scales - YouTube
6. Calculate scale scores: If you have determined that your items are valid and reliable, you
can calculate scale scores. To create composite scores for each scale, you can use SPSS
syntax or formulas. For example, you can create a new variable representing the sum or
average of individual items for each scale. To do this, go to "Transform" > "Compute
Variable". In the "Compute Variable" window, provide a name for the new variable, enter the
formula (e.g., SUM (item1 to item5) or MEAN (item1 to item5), and click "OK".
7. Perform further analyses: Once you have established the validity of your variables and
calculated scale scores, you can use them in other statistical analyses, such as regression,
ANOVA, or correlation, depending on your research question and objectives.
It is important to interpret the validity analysis results in the context of your research and
consider any specific requirements or guidelines for your study. If you need further guidance
or assistance, consult SPSS 16 Help menus or refer to statistical textbooks for more detailed
instructions and explanations
3. RELIABILITY ANALYSIS
a. Basic Reporting
Table 4 : Cronbach's Alpha
Subscale No of Item Cronbach's Alpha
Technology Acceptance 3 .829
Technology Reliability 3 .762
Perceived Confidence 5 .752
Cost Benefits 3 .840
Facilitation 5 .876
Sample Explanation
According to Hair (2016), a general accepted rule is that, Cronbach's Alpha of 0.6-0.7
indicates an acceptable level of reliability, and 0.8 or greater a very good level. Based on the
presented result in Table 1, the technology acceptance, technology reliability, perceived
confidence, cost benefits and facilitations have a good internal consistency, with a Cronbach
alpha coefficient reported of .829, .762, .752, .840 and .870 respectively. The results indicates
that all items used in the survey questionnaire are reliable.
4. CROSS-TABULATION ANALYSIS
(Non-parametric test – if the distribution of data is not normal)
a.) Basic Reporting
Table 4 (a) : GENDER * QUALIFICATION Cross-tabulation
QUALIFICATION
Female Count 5 6 23 20 14 68
Sample Explanation
You also need to explain Table 4(a) accordingly. You can support your explanation using
graph which can also be set in SPSS.
b. Advanced Reporting
Table 4(b) Chi-Square Tests
Pearson Chi-Square (Gender) Asymp. Sig df
Technology Acceptance .403 8
Qualification .366 4
a. 1 cells (10.0%) have expected count less than 5. The minimum expected count is 4.99.
Table 4(b) indicates the results of the square test (Pearson Chi Square) in the table above (c₂
= .403, df = 4, p> .05) showed no difference between Gender and Technology Acceptance
and between gender and qualification (c₂ = .366, df = 4, p> .05). This means that the
differences between the genders of men and women do not directly affect the qualification.
Students’ Notes
c₂ à This symbol is used for Chi-Square
5. CORRELATION ANALYSIS
Sample Explanation
a. Basic Reporting - Bivariate Correlation (Pearson Correlation)
Results of the Pearson correlation indicates that there is a moderate significant relationship
between technology acceptance and facilitation r=.469, p<.05 (n=150). It is also found that
the relationship between technology acceptance and perceived confidence is moderately
significant r=.425, p<.05 (n=150). However, it is found that there is a weak negative
relationship between technology acceptance and cost benefits r = -.247, p<.05 (n=150) which
indicate that an increase in the cost benefits will decrease technology acceptance.
b. Advanced Reporting
Correlation analysis is used to describe the strength and direction of the linear relationship
between two variables. Cohen (1988) suggests the following guidelines to determine the
strength of the relationship between variables as follows:
Results of the Pearson correlation indicates that there is a moderate significant relationship
between technology acceptance and facilitation r=.469, p<.05 (n=150). It is also found that
the relationship between technology acceptance and perceived confidence is moderately
significant r=.425, p<.05 (n=150). However, it is found that there is a weak negative
relationship between technology acceptance and cost benefits r = -.247, p<.05 (n=150) which
indicate that an increase in technology acceptance will decrease the cost benefits.
Students’ Notes
Below are the table generated from SPSS. You need to transform the table into more precise
data to explain the statistical results.
Correlations
Standardize
Model B Error Beta t Sig. Bound Bound order Partial Part Tolerance VIF
PERCON .305 .074 .313 4.116 .000 .159 .451 .425 .322 .283 .813 1.230
COSTBEN -.005 .057 -.007 -.086 .931 -.117 .107 -.274 -.007 -.006 .764 1.309
FACL .366 .073 .376 5.012 .000 .222 .510 .469 .383 .344 .838 1.194
a. Dependent Variable:
TECHACCEPT
he effect of
les (IV) on
(DV).
significant
and DV.
e larger than
en.
Model Summaryb
Change Statistics
on
8 indicates a
relation
Student Notes
The above tables are the data copied from SPSS output. We must choose the number that we
need only to report on the thesis/assignment. Based on the report above, we need to follow
the steps as listed below.
Step 1 : Report on Multicollinearity
Step 2 : Evaluating the Model including normality data using PP Plot
In multiple regression analysis, PP Plot can also tell you whether your data is normal or not.
A normal distribution of data indicates your model is fit.
Why is Model Fitting Important? As previously mentioned, a well-fit model does not match
every data point given but follows the overall curves. This shows that the model is neither
underfit nor overfit. If a model is not properly fit, it will produce incorrect insights and should
not be used for making decisions.
Notes to student
A linear regression analysis is used to predict changes in the dependent variable based on the
value of independent variables. Below is the standard APA format of reporting for linear
regression.
A linear regression was calculated to predict [dependent variable] based on [independent
variable] . A significant regression equation was found (F(_,__)= __.___, p < .___).
For further learning, you can also refer to V14.16 - Testing Mediation in SPSS - Example 1 -
YouTube and Introduction to Mediation Analysis | University of Virginia Library Research
Data Services + Sciences.
M
a b
X c Y
Step 1 : Estimate the total effect between X (IV) and Y (DV) using regression;
Step 2 : Estimate direct effect between X and M using bivariate regressions;
Step 3a: Estimate direct effect between X and Y using multiple regression with X and
M as predictors and Y as dependent variable.
Step 3b: Estimate direct effect between M and Y using multiple regression with X and
M as predictors and Y as dependent variable. This test is done to estimate co-
efficient and standard errors to estimate indirect effect for statistical
significance.
Step 4 : Estimate and test indirect effect for statistical significance using SOBEL test.
Based on data provided, a researcher hypothesizes that the effect between Self-Efficacy and
Intention may be mediated by Level of Integrity.
Hence;
X = Self Efficacy
Y = Intention to Engage
M = Level of Integrity
Level of Integrity
a b
X c Y
X Y
Coefficientsa
Lower Upper
Model B Std. Error Beta t Sig. Bound Bound
Self-efficacy
Lower Upper
Model B Std. Error Beta t Sig. Bound Bound
Mediation result
a=.737 M
(0.082)
b
c
X Y
Coefficientsa
Lower Upper
Model B Std. Error Beta t Sig. Bound Bound
a=.737 M
(.082)
b=.227
(.081)
X Y
c =.778
(.079)
Self-efficacy (IV) Intention (DV)
So, now we have all the figure to estimate indirect effect à a*b=Indirect effect. We will test
indirect effect for statistically significant.
It is also observed that direct effect between Self-Efficacy is statistically significant with
Integrity where p<0.05.
Step 4 : Use SOBEL test in order to test indirect effect for statistical significant.
4. Observe Sobel Test result. We look at p-value for indirect effect is 0.00746341 or
p<0.05. Result of Sobel Test shows indirect effect of M (Integrity) is statistically
significant. Therefore, we conclude that, Integrity (MV) indirectly mediate the
relationship between Self-Efficacy (IV) and Intention (DV).
5. Summary of the Statistical Result – This table should be presented all in the last
subheading for all variables.
Level of Integrity
a=.737 M
(.082)
b=.227
(.081)
X Y
c =.778
(.079)
Self-efficacy Intention
Steps
i. Observe the output table from SPSS :-
ii. Direct X àY (Model 1)
iii. Indirect Effect X àMàY (Model 2)
iv. Get all the figures in the table
v. Explain in APA format
You can refer to this picture for guide (enlarge it for clearer view)):-
How to report multiple regression with mediating variables? Please refer to Table 1 for
guidance. You also must open the SPSS output for figure reference (File name - Sample
Data with Mediation uum.spv).
Table 1 : Regression Analysis for Mediation of Level of Integrity between Self-Efficacy and
Intention.
Table 1 shows the impact of Self-Efficacy and Level of Integrity on Intention for
academicians in Private Universities in Klang Valley. In model 1, the R2 value of .83 reveals
that the Self-Efficacy explained 83% variance of intention with F(1,62)=296.8, p<0.001). The
findings revealed that Self-Efficacy positively predicted intention (β = .91, p<001. In model
2, the R2 of .85 reveals that the Self-Efficacy and Level of Integrity explained 85% variance
in the Intention with F (2, 61) = 169, p<0.001). The finding reveals that self-efficacy (β = .75,
p<001) and Level of Integrity positively predicted Intention (β = .21). The ΔR2 value of .85
reveals 85% in the variance of model 1 and model 2 with ΔF (2, 61)=168.9, p<0.001. The
regression weight for self-efficacy subsequently reduced from mode 1 to model 2 (.91 to .75)
but still remained significant which confirmed the partial mediation. More specifically, Self-
Efficacy has direct as well as indirect effect on intention.