You are on page 1of 2

Partial Correlation Interpretation: This sample does not provide

When one independent variable and one evidence of difference in the mean skin fold
dependent variable are studied by keeping other Example thickness of the three groups.
variables constant that is effect of other variables The correlation coefficients between serum
neutralized is called partial correlation. cholesterol(y), weight (x) and systolic blood ANOVA: Assumption:
 Partial correlation estimates the relationship pressure(z) among 15 normal male adults are Total variation present in a set of data is 1. The observations are independent and
between two variables while removing the given below partitioned into two or more components. randomly selected sample values.
influence of a third variable from the rxy =.402, ryz=.642, rxz=.932 Associated with each of these components is a 2. Population values are normally distributed
relationship. For example, a researcher
might want to know the relationship
Using the simple correlation coefficients, the
partial correlations are:
specific source of variation. Analysis makes it
possible to ascertain the magnitude of the
3. Population variances are equal. e.g. σ 12= σ 22
2
between height and weight. However, she is r xy −r xz r yz contribution of each of these sources to the total = σ3
aware that people’s bone and muscle r xy . z = 2 2
variation.
structures vary according to gender. She can
calculate a partial correlation between
√1−r √1−r
xz yz
skin fold thickness (mm) of three occupation
groups
height and weight while removing (holding 0.402−(.932∗.642)
=
constant, eliminating) the effect of gender
on the correlation.
√(1−.868)(1−.412) Total Sum of Square= ∑ x
( ∑ x )2 = 2
Q. Why ANOVA – Why not multiple t Tests?
−0.196 − If 5 treatments (e.g. a, b, c, d & e) are compared
 Partial correlation measures the strength of = = -.705 N for their effects, the possible number of pairs of
a relationship between two variables, while .279
controlling for the effect of one or more 2
4 +6 +8 +2 +6 +7 +3 +4 +5 -
(45)2 = 255-
2 2 2 2 2 2 2 2
comparison are: ab, ac,…… = 10. so, there will be
10 t tests. If each test is done at 5% level of
other variables. Figure. Proposed Three Variable Model 9 significance =.05), the probability of failing to
Assumptions 225=30 reject a hypothesis is = 0.95. P (accept Ho) for Pair
1. Useful in only small models like the models
Between gr. SS = 1, Pair 2 ….….. Pair 10 = (.95) 10 = .5987.Therefore, P
which involve three or four variables.
(∑ X 1)2 (∑ X 2)2 (∑ X 3)2 ( ∑ X )2
( reject Ho) for all 10 pairs =1-0.5987 =0.4013.So
2. Used in only those models which assume a
linear relationship. + + − overall  =0.4013.
3. The data is supposed to be interval in nature. n1 n2 n3 N  Decrease level of confidence
4. The residual variables or unmeasured 2 2 2 2  Increase significance error
variables are not correlated with any of the 18 15 12 – 45 = 231 -225 = 6 Q. Multiple Correlations
= + +
variables in the model, except for the one for 3 3 3 9 The multiple correlation coefficient is symbolized
which these residuals have occurred. Within gr. SS= Total SS – between gr. SS =30 -6 = by R and shows the correlation among more than
In partial correlation researcher need to have: 24 two variables. The multiple correlation ranges
1. Two variables, one continuous (e.g. test Applying the formula, we see that the partial from 0(when the independent variables in no way
score) and other is dichotomous (e.g. correlation between income and support for a help to predict Y) to 1 (When the independent

SS

MS
df

W) ratio (MS
groupBetween of Source

Bet. /MS Variance


Gender) decreased role for government in the economic variables predict Y with complete accuracy).

P
2. 1-3 antecedent or interviewing variables. arena is 0.02 For example, if variables 1,2,3 were being
Partial Correlation formula correlated and variable 1 were the dependent
r xy −r xz r yz
Day laborer 1

Business 2

Service 3

3 variable and 2,3 were independent variables, the

3
r xy . z= (where X and Y = = multiple correlation coefficient would be written
2
√1−r √1−r
xz
2
yz
4 R1.23.
groupWithin

P > 0.05
0.75 Thus, for three variables designated X1, X2 and X3,

24
independent and dependent variables and Z =

4
control variable.) R1.23 is a measure of the degree of association
between dependent variable X1 and the
Total

Where ryx.z  = the (first-order) partial correlation of


4 2 3 independent variable X2 and X3 jointly. The most
30
y and x controlling for z
8

-
ryx = the bivariate (zero-order) correlation of y and 6 6 4 convenient form of expressing the multiple
x 8 7 5 correlation coefficient in terms of a simple
ryz = the bivariate (zero-order) correlation of y and Tabulated value of F (2,6) at 5% level F tab=5.143 > correlation coefficients for 3- variable case(sutra)
Total 18 15 12 45
z Fcal 0.75: So, P >. 05. Example. Given the following values of X1, X2 and
rxz = the bivariate (zero-order) correlation of x and
Summation ∑X1 ∑X2 ∑X3 ∑X Conclusion: Not significance. Do Not Reject Ho. X3. Compute the multiple correlation coefficient
of X
z R1.23.
N 3 3 3 9
1

n
2

n
3

N
Solution To make use of the formula, we first First, you must find the residuals. procedure prolongs life more? And so, survival
compute the following simple correlation time is the response variable. The type of therapy
coefficients: r12, r13 and r23, where r12, r13 and r23 are given is the explanatory variable; it may or may
respectively the coefficients of correlation not affect the response variable. In this example,
between the pairs (x1, x2), (x1, x3) and (x2, x3). we have only one explanatory variable: type of
It can be shown that r12=.768 r13=.965 and r23=.812 treatment. In real life we would have several more

2 r 212+r 213−2r 12 r 13 r 23 explanatory variables, including: age, health,


R 1.23
= weight and other lifestyle factors.
1−r 223 Q. R2 or r2 (coefficient of determination)
In statistics, the coefficient of determination,
=
denoted R2 or r2 and pronounced "R squared", is
.768 + .9652−2(.768)(.965)(.812)
2
the proportion of the variance in the dependent
variable that is predictable from the independent
1−.8122 variable(s)
=.93 The coefficient of determination is a more
Hence R1.23= .96 powerful measure of strength of the relationship
Limits of R between x and y variable. It is presented as a
The multiple correlation coefficient R is always proportion or as a percentage.
non-negative. Symbolically, 0<R<1.
Explained variation
R2 =
Q. Residual Total variation
In regression analysis, residual is the difference Suppose r2 = .40, it means 40% of the variation in y
between the observed y-value (from scatter plot) Compute: scatter plot y-value minus regression is explained by the variation in x.
and the predicted y-value (sometimes referred to line y-value for each point. The Coefficient of determination can be negative,
as "observed error"). Each data point has one D1 = 2 - 1.4 = 0.6 although this usually means that our model is a
residual. D2 = 1 - 2.1 = -1.1 poor fit for your data. It can also become negative
Residual = Observed value - Predicted value D3 = 3.5 - 2.8 = 0.7 if you didn’t set an intercept.
e=y-ŷ D4 = 3 - 3.5 = -0.5
Both the sum and the mean of the residuals are D5 = 4.5 - 4.2 = 0.3
equal to zero. That is, Σe = 0 and e = 0. Now, plot the residuals.
Residuals help to determine if a curve (shape) is (1,0.6), (2, -1.1), (3,0.7), (4,-0.5), (5,0.3)
appropriate for the data. (linear versus non-linear)
Example:
y = 0.7x + 0.7.

Q. Response variable
A response variable is the variable about which a
researcher is asking a question. He or she wants to
know if this variable 'responds' to other factors
being examined.
The response variable is also known as the
dependent variable. In other words, the response
variable depends on another factor, known as the
independent(explanatory) variable, to cause
change, or response.
Example: we’re trying to figure out if chemo or
anti-estrogen treatment is better procedure for
breast cancer patients. The question is: which

You might also like