You are on page 1of 19

A peer-reviewed electronic journal.

Copyright is retained by the first or sole author, who grants right of first publication to the Practical Assessment, Research & Evaluation.
Permission is granted to distribute this article for nonprofit, educational purposes if it is copied in its entirety and the journal is credited.
PARE has the right to authorize third party reproduction of this article in print, electronic and database forms.
Volume 17, Number 9, April 2012 ISSN 1531-7714

Interpreting Multiple Linear Regression:


A Guidebook of Variable Importance
Laura L. Nathans, University of North Texas
Frederick L. Oswald, Rice University
Kim Nimon, University of North Texas

Multiple regression (MR) analyses are commonly employed in social science fields. It is also
common for interpretation of results to typically reflect overreliance on beta weights (cf. Courville &
Thompson, 2001; Nimon, Roberts, & Gavrilova, 2010; Zientek, Capraro, & Capraro, 2008), often
resulting in very limited interpretations of variable importance. It appears that few researchers
employ other methods to obtain a fuller understanding of what and how independent variables
contribute to a regression equation. Thus, this paper presents a guidebook of variable importance
measures that inform MR results, linking measures to a theoretical framework that demonstrates the
complementary roles they play when interpreting regression findings. We also provide a data-driven
example of how to publish MR results that demonstrates how to present a more complete picture of
the contributions variables make to a regression equation. We end with several recommendations
for practice regarding how to integrate multiple variable importance measures into MR analyses.

Across behavioral science disciplines, multiple generally, associations when there are correlations
linear regression (MR) is a standard statistical technique between independent variables. The more predictors
in a researchers toolbox. An extension of simple linear there are in the model, the greater the potential there is
regression, MR allows researchers to answer questions for multicollinearity or association between variables
that consider the role(s) that multiple independent (Pedhazur, 1997; Zientek & Thompson, 2006). The
variables play in accounting for variance in a single current paper demonstrates that there are several
dependent variable. Researchers tend to rely heavily on approaches researchers can use to gain insight into the
beta weights when interpreting MR results (e.g., role(s) that predictors play in MR that are particularly
Nimon, Gavrilova, & Roberts, 2010; Zientek, Carpraro, important to use in the presence of associations or
& Capraro, 2008). Presumably, this is because most correlations between variables. Each approach, we
statistical packages automatically output these weights show, yields different perspectives and insights
by default, which can then be easily rank ordered based regarding the importance of independent variables in a
on their magnitudes. But beta weights serve as only regression equation, as well as often different rank
one method for answering the proverbial fairy-tale orderings of those independent variables in terms of
question: Mirror, mirror on the wall, what is the best predictor their contributions to the regression equation. We
of them all? As others have shown (e.g., Courville & support Azen and Budescus (2003) assertion that
Thompson, 2001) and readers will see, it is often not assessment of variable importance is contingent on how
best to rely only on beta weights when interpreting MR importance is defined and quantified and that it is
results. In MR applications, independent variables are therefore impossible to generate a, precise universal
often intercorrelated, resulting in a statistical definition of importance (Budescu, 1993, p. 544). The
phenomenon that is referred to as multicollinearity when term variable importance in and of itself is not a
correlations between predictors are high, and, more meaningful term; rather, it must be discussed (if it is
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 2
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

discussed at all) in the context of a specific metric for it an independent and a dependent variable that does not
to have any meaning to the researcher or the reader. It incorporate variance contributions of other
is probably better for researchers to emphasize the independent variables, as opposed to a direct effect in a
specific ways in which variable importance is path model, which refers to an independent variable
operationalized (e.g., dominance analysis, commonality that directly impacts another variable in the model
analysis) rather than to focus on the blanket term and (either an independent or dependent variable).
assume that the reader knows what this term means. Different techniques for assessing variable importance
Through gaining an understanding of multiple methods from these different categories may potentially (but not
of assessing variable importance and how they in all cases) yield different rank orderings of
complement each other, a researcher should be able to independent variables. Use of multiple techniques that
avoid dichotomous thinking (e.g., yes it is an reflect these three types of effects when assessing
important predictor, or no it is not), and instead variable importance will yield the richest and most
understand the importance of independent variables in complete picture regarding the relationships between
more nuanced terms. This means that there is no single independent variables and the dependent variable.
right way to interpret regression results, and although Keeping this context in mind, the goal of our
reliance on beta weights may feel right because it is paper is to present a set of measures or lenses that
normative practice, it provides very limited provide different yet complementary ways of viewing
information. We want to open researchers eyes to the the role(s) that each independent variable plays in a MR
multiple lenses through which MR can be profitably equation. Our paper is structured by (a) defining each
viewed; we also want researchers to promote the value measure, (b) highlighting each measures advantages
of the multiple-lens approach in their publications and and limitations, (c) describing how each measure can
with their colleagues and graduate students. help identify suppression effects, (d) outlining specific
Several statistical techniques have been developed research questions that each measure can address, and
to determine independent variables contributions to (e) detailing when the researcher should select each
MR models. Our focus is on two general families of particular index to assess variable importance. This
techniques. One family provides different methods of guidebook should serve as a practical resource for
rank ordering individual predictors contributions to an researchers to define and identify the measure or set of
overall regression effect or R2 (e.g., Pratts measure, measures with which to analyze their data using MR.
dominance analysis, relative weights), and the other In conclusion, we present a data-driven example and
family involves partitioning R2 into the unique and results section that researchers can use as a template for
shared variance contributions of the independent interpreting and reporting MR results. Lastly, we
variables (e.g., commonality analysis, squared semi- present recommendations for practice for selecting and
partial correlations). These two families are aligned reporting of the variable importance measures included
with the framework of LeBreton, Ployhart, and Ladd in our guidebook.
(2004), who categorized methods of variable importance
Beta Weight
into those that assess (a) direct effects, which quantify the
contribution of each independent variable to the Research shows that beta weights are heavily relied
regression equation when measured in isolation from on to assess variable importance (e.g., Courville &
other independent variables; (b) total effects, which Thompson, 2001; Nimon, Gavrilova, & Roberts, 2010;
quantify the each independent variables contribution Zientek, Carpraro, & Capraro, 2008). The regression
to the regression equation when the variance weight for each given independent variable is
contributions of all other predictors in the regression interpreted as the expected difference in the dependent
model have been accounted for; or (c) partial effects, variable score between people who differ by one unit
which quantify the each independent variables on that independent variable, with all other
contribution to the regression equation while independent variable scores held constant (Hoyt,
accounting for the contributions to regression models Leierer, & Millington, 2006; Johnson, 2004). When
of a specific subset or subsets of remaining variables are standardized (i.e., converted to z-scores),
independent variables. It is important to clarify that a regression weights are called beta weights. A beta
direct effect in this paper refers to a relationship between weight for an independent variable indicates the
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 3
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

expected increase or decrease in the dependent associations between independent variables. A given
variable, in standard deviation units, given a one beta weight may receive the credit for explained
standard-deviation increase in independent variable variance that it shares with one or more independent
with all other independent variables held constant. variables (Pedhazur, 1997). As such, the other weights
When variables are not standardized (i.e., scaled in their are not given credit for this shared variance, and their
original metric), regression weights are called B weights. contribution to the regression effect is thus not fully
A B weight also indicates how much a one unit increase captured in the beta weight value.
in the independent variable results in an increase in the Beta weights are also limited in their ability to
dependent variable with all other variables held determine suppression in a regression equation. An
constant. However, this increase is scaled in terms of independent variable that contributes little or no
the variables original scaling metric, rather than in a variance to the dependent variable may have a large
standardized metric. Our focus in this paper is on beta non-zero beta weight because it purifies one or more
weights rather than B weights, because beta weights are independent variables of their irrelevant variance,
more comparable across independent variables due to thereby allowing it or their predictive power to increase
being scaled in the same standardized metric. (Capraro & Capraro, 2001). In such a case, the beta
According to Pedhazur (1997), beta weights are weights value may lead the researcher to erroneously
computed to weight the independent variables so that conclude that it directly predicts the dependent variable
when the weights are multiplied by variable scores, (Pedhazur, 1997). Thus, after obtaining beta weights,
their sum is maximally correlated with the dependent researchers should not cease to analyze their MR
variable. This computation process minimizes the sum results, but rather should consider additional measures
of squared errors between the dependent variable to gain a broader and fuller perspective on the
scores and the dependent variable scores predicted by contributions that independent variables make to the
the regression equation (called Y-hat or ) (Pedhazur, regression equation.
1997). Because the beta weight calculation process Beta weights are best used as an initial starting
accounts for the contributions of all variables in the point from which to begin exploring the issue of
model to the regression equation, each beta weight is a independent variables contributions to a regression
measure of the total effect of an independent variable equation. It is recommended that all researchers begin
(cf., LeBreton, Ployhart, & Ladd, 2004). If beta MR analyses with beta weights, as they are easily
weights are rank-ordered by their absolute values, it is computed with most statistical software packages and
important to understand that the rank ordering can provide an initial rank ordering of variable
represents solely this type of contribution (and not contributions in one computation. If there are no
others that we will describe shortly). associations between independent variables or the
According to Pedhazur (1997), sole reliance on model is perfectly specified (Courville & Thompson,
using beta weights to interpret MR is only justified in 2001), no other techniques need to be employed.
the case where predictors are perfectly uncorrelated. In However, in the more realistic case of correlated
the absence of shared variances between independent predictors, researchers should select from at least one
variables, each standardized beta weight is equal to the of the other techniques discussed in this guidebook to
zero-order correlation between the independent and determine the impact of shared variance between
dependent variable. In such a case, variable importance variables on the regression equation.
can easily by determined by squaring the standardized Beta Weight Research Question: What is the
beta weights; thus, it is not necessary to calculate more contribution of each independent variable to the regression
complicated MR indices. equation, holding all other independent variables constant?
The major advantage of beta weights is that they Zero-Order Correlation
provide a measure of variable importance that is easily
computed and provides an initial rank ordering of As applied to MR, zero-order correlations reflect
independent variables contributions to a MR equation the bivariate relationships between independent and
that accounts for contributions of other independent dependent variables. According to Hinkle, Wiersma,
variables. However, this lens has a limited focus due to and Jurs (2003), the correlation coefficient is, an index
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 4
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

that describes the extent to which two variables are s/he should choose the variable with the highest zero-
related (p. 98). The correlation coefficient reflects order correlation. Second, zero-order correlations are
both the magnitude and direction of the relationship less sensitive to effects of sampling error than are beta
between two independent variables. Its values range weights. Third, this measure is the only measure
from -1.0 to 1.0. If a correlation coefficient is negative, presented in this guidebook that is able to quantify how
the values of the two variables that are correlated are much variance is directly shared between the
inversely related; as one variables scores increase, the independent and dependent variable without being
other variables scores decrease. If a correlation affected by shared variance between independent
coefficient is positive, an increase (or decrease) in one variables.
variable is related to an increase (or decrease) in the According to Nunnally and Bernstein (1994), there
other variable in the coefficient. The closer the value are limitations to the use of zero-order correlations.
of the correlation coefficient is an absolute value of 1.0, First, an independent variable may have the largest
the larger the magnitude of the relationship is between zero-order correlation with the dependent variable yet
the two variables. If the value of the correlation make the smallest (or potentially no) contribution to
coefficient is zero, there is no relationship between the the regression equation when measures of total effect
two variables. It is important to note that the are calculated due to variance it shares with other
strength/magnitude of a correlational relationship is variables. As such, several independent variables may
not related to the sign of the correlation coefficient; show high zero-order correlations, yet may not be,
thus, equivalent predictive power can be attributed to particularly important to independent prediction (p.
correlations of equivalent magnitude but different 192) when other independent variables contributions
signs. to the regression equation are accounted for. Thus, it
Zero-order correlations are measures of direct is best to complement use of zero-order correlations
effect (cf., LeBreton, Ployhart, & Ladd, 2004), as they with measures of total and partial effects that consider
determine the magnitude of the bivariate relationship contributions of other variables to the regression
between the independent and dependent variable equation in their assessments of variable importance.
without accounting for the contributions of other Zero-order correlations are often useful to identify
variables in the regression equation. These coefficients the occurrence of suppression when viewed in concert
are also commonly referred to as validities (Nunnally with beta weights, although they cannot identify
& Bernstein, 1994). In the case where independent suppression effects in and of themselves. If an
variables are uncorrelated, zero-order correlations are independent variable has a near-zero or negligible zero-
equivalent to beta weights and are sufficient to rank order correlation with the dependent variable and a
order independent variables. When squared, they add large and statistically significant beta weight, these two
up to the model R2, and thus partition the regression statistics suggest that the variable is a suppressor. This
effect (Pedhazur, 1997). Conversely, in the case where variable shares no variance directly with the dependent
at least some of the independent variables are variable and thus contributes to the regression equation
correlated, squared zero-order correlations will through removing irrelevant variance from other
generally add up to a total that is greater than the independent variables.
model R2, as shared variance in the dependent variable
is added multiple times into the overall sum for R2. Researchers should choose this measure as a
Thus, in the latter case, researchers should employ complement to beta weights when they are interested in
other statistics to determine how the regression determining the magnitude and direction of the
equation is affected by shared variance among the bivariate relationship between an independent variable
independent variables. and the dependent variable without accounting for
other predictors in the regression equation.
According to Nunnally and Bernstein (1994), there
are several advantages to use of zero-order correlations. Zero-Order Correlation Research Question: What is
First, if a researcher has to select one independent the magnitude and direction of the relationship between the
variable to target for research or intervention after independent and dependent variable, without considering any
obtaining several beta weights of similar magnitude, other independent variables in the MR model?
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 5
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Product Measure weights to provide a different perspective on variable


importance because its values, unlike those for beta
Pratt (1987) proposed the product measure, which
weights, do sum to R2. It can be compared and
is calculated by multiplying the variables zero order
contrasted with other methods for partitioning the
correlation (its relationship to the dependent variable in
regression effect as well.
isolation from of other independent variables) by its
beta weight (which accounts for contributions of all Product Measure Research Question: How can the
other predictors to the regression equation). Thus, the regression effect be partitioned into non-overlapping partitions
product measure uniquely reflects in one statistic both based on the interaction between the beta weight of each
direct and total effects (cf. LeBreton, Ployart, & Ladd, independent variable and its zero-order correlation with the
2004). dependent variable?
A benefit of product measures is that they Relative Weights
partition the regression effect; thus, their sum across all Another method for partitioning the R2 in MR
independent variables equals the multiple R for the between independent variables in the model is through
regression model, even when independent variables are relative weight analysis (RWA; Fabbris, 1980; Genizi, 1993;
correlated with one another (Azen & Budescu, 2003). Johnson, 2000). Johnson (2000) explained that when
This measure thus enables rank orderings of variable independent variables are uncorrelated, the relative
importance based on the partitioning of the regression weights are computed by calculating the squared zero-
effect. The major difficulty with this measure is that it order correlation between the independent variable and
may also yield negative values for an independent the dependent variable (also the standardized beta
variable even when the independent variable accounts weight) and dividing this number by R. In contrast,
for a large amount of variance in the dependent when independent variables are correlated, relative
variable, particularly if either the zero-order correlation weights address this problem by using principal
or beta weight for an independent variable is negative components analysis to transform the original
(Darlington, 1968). If large negative values are independent variables into a set of uncorrelated
generated for independent variables, it renders their principal components that are most highly correlated
product measure values substantively meaningless in with the original independent variables (Tonidandel &
terms of quantifying their contribution to R2 because LeBreton, 2010). These components are then
their variance contributions to the regression effect are submitted to two regression analyses. The first analysis
subtracted from rather than added to the overall R2. is a regression that predicts the dependent variable
Thus, they should not be used in such cases. from these uncorrelated principal components. Next,
Concerning suppression, Thomas, Hughes, and the original independent variables are regressed onto
Zumbo (1998) explained that the product measure can the uncorrelated principal components. Finally, relative
identify suppressor variables. Multiplying an weights are computed by multiplying squared
independent variables beta weight by the small or regression weights from the first analysis (regression of
negligible zero-order correlation of the suppressor dependent variables on components) with squared
variable with the dependent variable will yield a small regression weights the second analysis (regression of
or negligible product measure value, thereby independent variables on components). Each weight
demonstrating that the variable did not directly can be divided by R2 and multiplied by 100 so that the
contribute to the regression effect. Small negative new weights add to 100%, with each weight reflecting
values may also indicate the presence of suppression if the percentage of predictable variance. Relative
a near-zero correlation coefficient is multiplied by a weights are unique as a measure of total effect in that they
negative beta weight (a classic profile for suppression). provide rank orderings of individual independent
variables contributions to a MR effect in the presence
This measure should be selected as an initial, easily
of all other predictors based on a computational
computed (as opposed to general dominance and
method that addresses associations between
relative weights, which will be discussed later) method
independent variables between variables by creating
of partitioning R2 in the presence of correlated
their uncorrelated counterparts (Johnson &
predictors. It should be used to complement beta
LeBreton, 2004).
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 6
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

There are several strengths of relative weights. regression effect than other R2 partitioning methods
First, relative weights add up to R (Tonidandel & (such as the product measure).
LeBreton, 2011). They partition the regression effect Relative Weights Research Question: How do
based on a procedure that addresses the problem of independent variables contribute to the dependent variable when
associations between independent variables through the
the regression effect is partitioned as a joint function of (a) how
use of uncorrelated principal components. Relative highly related independent variables are to their uncorrelated
weights will thus, perform appropriately and much counterpart and (b) how highly related the uncorrelated
better than regression weights when partitioning counterparts are to the dependent variable?
variance in the presence of correlated independent
variables (Tonidandel & LeBreton, 2011, p. 5). Structure Coefficients
Additionally, relative weights are easy to explain to A structure coefficient is the bivariate correlation
researchers, because, unlike beta weights, they account between an independent variable variable and the
for all variance explained by the regression model and predicted value resulting from the MR model, where
at the same time address the minimize the impact of
represents the predicted dependent variable scores
multicollinearity on weights attributed to each
(Courville & Thompson, 2001). A structure coefficient
independent variable (Johnson & LeBreton, 2004).
in MR analyses is a useful measure of a variables direct
These points are probably more important than
effect (LeBreton, Ployhart, & Ladd, 2004), as it
another advertised benefit of RWA, which is
quantifies the magnitude of the bivariate relationship
computational ease (Johnson, 2000).
between each independent variable and in isolation
According to Tonidandel and LeBreton (2011), from other independent variable- correlations.
there are also several limitations of relative weight However, it is important to note that other
analysis. First, these weights are highly dependent independent variables do contribute indirectly to
upon the independent variables in the regression structure coefficient values, as they are used in
model, and are thus susceptible to model computation of the scores. The major difference
misspecification (p. 5). Second, although the between a zero-order correlation and a structure
computational procedure for relative weights presents a coefficient is that the structure coefficient is scaled to
unique way of dealing with associations between remove the difference of the multiple R2.
independent variables, it is a common myth (p. 5)
among users of this method that this method fixes all According to Courville and Thompson (2001),
related problems. Relative weights are still affected by there are two ways a structure coefficient can be
associations between independent variables; as each computed. First, for a given independent variable, X,
weight generally contains variance that is shared with the structure coefficient is:
other independent variables as well as unique variance,
associations between independent variables are not
completely eliminated. We would also add as a where rX.Y is the bivariate correlation between the
potential limitation that relative weights only identify independent variable (X) and the dependent variable
suppression effects indirectly, when the weights sum to (Y), and R is the multiple correlation for the regression
a total that is larger than R or account for greater than containing all independent variables. Second, the
100% of the variance in R when relative weights are structure coefficient may be calculated by computing
converted to percentages. the correlation between a given independent variable
and the predicted scores, or:
The researcher should generally select relative
weights when there is multicollinearity between
independent variables. It serves as a strong
complement to beta weights as, it is uniquely able to When squared, structure coefficients represent the
partition R2 while minimizing the impact of amount of variance that an independent variable shares
associations between variables, and can thus present a with the variance from the predicted scores.
more accurate picture of variables contributions to a According to Courville and Thompson (2001), a
beneficial property of structure coefficients is that they
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 7
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

are not affected by associations between independent variance it shares with another independent variable to
variables, as a structure coefficient is simply a Pearson r another beta weight.
between an independent variable and . Thus, Squared Structure Coefficients Research Question:
structure coefficients enable an understanding of how How much variance in the predicted scores for the dependent
much variance each independent variable shares with variable ( ) can be attributed to each independent variable when
the predicted scores that are the actual analytic focus variance is allowed to be shared between independent variables?
of the study. If an independent variable has a small beta
weight but explains substantial variance in as Commonality Coefficients
reflected a large squared structure coefficient, the Developed in the 1960s as a method of
researcher then knows that (a) there is shared variance partitioning variance (Mayeske et. al, 1969; Mood, 1969,
between this variable and another variable and (b) that 1971; Newton & Spurrell, 1967), commonality analysis
the beta weight calculation process assigned that shared partitions the R that is explained by all independent
variance to another independent variable. However, variables in a MR into variance that is unique to each
structure coefficients are limited in and of themselves variable and variance that each possible subset of
due to being solely a measure of direct effect that does not independent variables share (Onwuegbuzie & Daniel,
identify which independent variables jointly share 2003; Rowell, 1996). For example, if a MR models the
variance in predicting the dependent variable or effects of X1, X2 and X3 in predicting Y (the dependent
quantify the amount of this shared variance. variable), then a commonality analysis calculates the
A special case that highlights the usefulness of amount of variance in Y that is predicted by 7 subsets
structure coefficients in identifying how the variance that together add up to the model R: the unique
assignment process for a particular regression equation variances of each variable X1, X2, and X3; the shared
occurs is the suppression case. If a structure coefficient variances of variables taken two at a time, {X1, X2},
is near zero or zero in magnitude, that independent {X1, X3}, and {X2, X3}; and the shared variance of all
variable contributes little or no direct variance to . If three variables {X1, X2, X3}. It is important to note
that independent variable has a substantial beta weight, that this process partitions the regression effect into
it can be determined that the particular independent nonoverlapping components of variance that can thus
variable is a suppressor. Notably, a similar process is be easily compared.
used when comparing zero-order correlations to beta There are two types of commonality coefficients:
weights to determine the presence of a suppression unique effects and common effects. Unique effects
effect. The major difference between the two processes reflect how much variance an independent variable
is that the independent variable in the zero-order contributes to a regression equation that is not shared
correlation in a suppression effect will share little or no with other independent variables (Zientek &
variance with the dependent variable, while the Thompson, 2006). This statistic is also termed the
independent variable in the structure coefficient will independent variables usefulness or squared
share little or no variance with the scores that are a semipartial correlation. If independent variables are all
portion of the dependent variable. However, the uncorrelated, all independent variable contributions are
researcher does not know what variables are being unique effects, as no variance is shared between
suppressed and must rely on other techniques, such as independent variables in predicting the dependent
commonality analysis (to be discussed), to determine variable. In this case, unique effects are identical in
the magnitude and loci of suppression. value to squared zero-order correlations and squared
As asserted by Courville and Thompson (2001) beta weights, and variable importance can be
and we recommend, the researcher should select this determined by rank orderings of unique effects. A
measure in addition to beta weights in the presence of unique effect is a measure of total effect, as it is only
correlated predictors, as it is able to determine both (a) calculated when all independent variables have been
entered into the regression equation.
the variance each independent variable shares with
and (b) if a variables contribution to the regression In contrast, common effects provide detailed
equation was distorted or minimized in the beta information that identifies and quantifies the extent and
weight calculation process due to assignment of pattern of the independent variables overlap in
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 8
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

predicting dependent variable variance (Mood, 1971). to report and interpret. However, software in PASW
Common effects are measures of total effect (LeBreton et (Nimon, 2010) and R (Nimon, Lewis, Kane, & Haynes,
al., 2004), as they quantify the contribution to the 2008) exists to compute these coefficients. Researchers
regression effect that each variable shares with every will also need to summarize multiple combinations of
other variable set. By knowing which variables share variables represented by commonality coefficients. For
variance in R, researchers have knowledge of how example, with six variables, there are second-order
particular sets of variables operate in combination in commonalities between two variables, third-order
predicting an outcome, and can thus generate commonalities between three variables, fourth-order
recommendations regarding how to jointly target these commonalities between four variables, fifth-order
variables to produce desired effects (see Seibold & commonalities between five variables, and a sixth-order
McPhee, 1979 for an example). Commonality commonality between all six of the variables (Amado,
coefficients sum to the multiple R for the regression 1999). It may be hard for the researcher to assign clear
model. Thus, they can be used to determine how the meanings and interpretations to higher-order
regression effect is partitioned into percentages of commonalities that reflect combinations of varied
unique and shared variance. constructs (DeVito, 1976), but we argue that the
attempt to do so can be worthwhile.
A unique property of commonality analysis is that
the researcher can add the common effects for each The results of a commonality analysis can aid in
independent variable and compare them with its unique identifying where suppressor effects occur and also
effect to determine whether a variable contributes more how much of the regression effect is due to
to a regression effect when operating in combination suppression. Negative values of commonalities
with other variables or independently of them. The generally indicate the presence of suppressor effects
resulting data provide rich interpretation of the (Amado, 1999). In the suppression case, a variable in a
regression effect. Seibold and McPhee (1979) stressed particular common effect coefficient that does not
the importance of decomposing R into constituent directly share variance with the dependent variable
parts: suppresses variance in at least one of the other
independent variables in that coefficient. The
Advancement of theory and the useful
suppressor removes the irrelevant variance in the other
application of research findings depend not
variable or variables in the common effect to increase
only on establishing that a relationship exists
the other variable(s) variance contributions to the
among independent variables and the
regression effect (DeVito, 1976; Zientek & Thompson,
dependent variable, but also upon determining
2006). Commonality analysis is uniquely able to both
the extent to which those independent
identify which variables are in a suppressor relationship
variables, singly and in all combinations, share
and the specific nature of that relationship. The
variance with the dependent variable. Only
researcher can look across a table of commonality
then can we fully know the relative importance
coefficients and see if a particular variable is a part of
of independent variables with regard to the
multiple negative common effects, which suggests that
dependent variable in question. (p. 355)
it is a suppressor for at least one of the other variables
in the common effects. Additionally, researchers can
There are two major difficulties with the
compare common effects with structure coefficients
commonality procedure. First, as the number of
and zero-order correlations to help identify
independent variables increases, the number of
suppressors. If a variable has a small or negligible
commonality coefficients increases exponentially
structure coefficient and zero-order correlation (and
(Mood, 1971). The number of commonality
coefficients is 2p-1, where p equals the number of thus does not directly share variance with or the
independent variables in the MR model (Mood, 1971). dependent variable) and that same variable is part of a
For example, with three, four, or five independent negative common effect, it can be determined that the
variables, the number of commonality coefficients is 7, variable is a suppressor for the other variables that are
15, and 31 respectively. Therefore, with large numbers part of this common effect. Summing all negative
of predictors, there are large numbers of commonalities common effects for a regression equation can quantify
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 9
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

the amount of suppression present in the regression dominates X1; (c) X1 and X2 contribute equally to
equation. prediction of variance in the dependent variable; or (d)
neither independent variable dominates the other
Researchers should select commonality analysis
across all possible model subsets (i.e., they each
when they have identified that shared variance between
dominate different model subsets or no subsets).
independent variables is impacting the regression
equation (usually through comparisons of beta weights Azen and Budescu (2003) developed a weaker type
with structure coefficients) so that they can determine of dominance, termed conditional dominance. Conditional
the patterns and extent of shared and unique variance. dominance is determined by first calculating the averages
Additionally, commonality analysis should always be of independent variables contributions to all models of
selected for use in the presence of suppression effects, the same subset size (e.g., how much unique variance,
as it is uniquely able to both identify the variables that on average, independent variables add to models with
are in a suppressor relationship and quantify the no independent variables, one other independent
amount of suppression present in a particular variable, and two other independent variables if the
regression equation. model has three independent variables). If, on average,
an independent variable contributes more unique
Commonality Analysis Research Question: How
variance than another independent variable across
much variance in the dependent variable is uniquely vs. jointly
model of all subset sizes, an independent variable is
explained by each predictor or predictor set?
said to conditionally dominate another independent
Dominance Weights variable.
Dominance analysis is a technique developed by Lastly, Azen and Budescu (2003) define a third and
Budescu (1993) and refined by Azen and Budescu weakest type of dominance, termed general dominance.
(2003) to determine variable importance based on General dominance reflects the average additional
comparisons of unique variance contributions of all unique variance contribution of each independent
pairs of variables to regression equations involving all variable to all subset models; this can be computed by
possible subsets of predictors. Budescu (1993) initially averaging across all conditional dominance statistics.
proposed dominance analysis as a means to improve Interpretively, general dominance represents the,
upon previous relative importance measures by average difference in fit between all subset models (of
quantifying an independent variables direct effect in equal size) that include Xi and those that do not include
isolation from other independent variables (as the it (Azen & Budescu, 2003, p. 137). An important
subset containing no other independent variables property of general dominance variance averages is that
includes squared zero-order correlations), total effect (as they partition the total R, enabling the researcher to
it compares independent variables unique variance rank order independent variables contributions to the
contributions when all predictors are included in the regression effect based on their average contributions
model), and partial effect (as it compares independent across all possible subsets of independent variables.
variables unique variance contributions for all possible Notably, this method of partitioning R2 is different
subsets of independent variables). from that used with both the product measure, which is
According to Azen and Budescu (2003), there are based upon zero-order correlations and beta weights
three types of dominance. First, an independent rather than averages of uniqueness values, and relative
variable shows complete dominance over another weights, which are calculated on the basis of the entire
independent variable across all submodels if the former regression model and not the average contribution
independent variable always shows a higher unique across all sub-models.
variance contribution than the latter independent There are many advantages of the dominance
variable when it is entered last into regression equations analysis procedure. First, dominance analysis enables
containing all possible subsets of independent variables. comparisons between the uniqueness values of all
Budescu (1993) lists four exclusive complete independent variables for all possible subsets of
dominance relationships that an independent variable independent variables in one technique (Budescu,
can have in relation to another independent variable: (a) 1993). Second, dominance analysis rank orders
X1 completely dominates X2; (b) X2 completely independent variables equivalently across multiple
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 10
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

measures of fit, including Mallows C, Akaikes to the regression equation through suppressing
information criterion, Bayesian information criterion, variance in other predictors. Viewing this trend for an
adjusted R, and many other measures (Azen & independent variable that contributes no or little
Budescu, 2003). Most importantly, dominance analysis variance to the subset containing no predictors (and
is a comprehensive technique in that it is the only thus is not a direct contributor to variance in the
measure that involves calculation of all three types of dependent variable) will demonstrate that the
effects in LeBreton et al. (2004)s theoretical independent variable is a classic suppressor.
framework. Researchers should select this technique when they
There are also several limitations of dominance are interested in understanding the dynamics of all
analysis in addition to its many advantages. First, possible subsets of independent variables, which is not
dominance analysis involves computation of numerous captured in any other measure in this guidebook. If the
statistics, and the number of models increases researcher wants to make pairwise comparisons
exponentially with the number of independent between variables, this technique is solely suited for
variables. For example, Johnson (2000) noted that those comparisons, as well. This technique is also well-
there are 1,023 models for 10 independent variables suited for selection of the subset of the most significant
and 32,767 models for 15 independent variables. The independent variables for future analyses.
researcher must construct tables of all Rs and perform Dominance Analysis Research Question: Does one
extensive visual comparisons to establish dominance independent variable contribute more variance that another
relationships for multiple pairs of variables. A SAS independent variable to the regression effect for models containing
macro is listed in Azen and Budescu (2003) to all or some subsets of independent variables, or on average across
automate the process, however. Second, Budescu all possible subsets of independent variables?
(1993) stressed that dominance analysis is contingent
upon identification of the correct regression model; The preceding discussions are summarized in Table 1.
rank orderings will less informative if important A Caveat Regarding Theory-Driven
variables are excluded and unimportant variables are Regression Methods and Stepwise
included. Third, Baltes, Parker, Young, Huff, and Regression
Altman (2004) did not find that in their data,
dominance analysis added new information to that Because of the importance of regression analysis
found with more traditional variable importance for answering theory-drivenas opposed to data-
measures except in the presence of independent drivenresearch questions, we provide a discussion of
variables that do not contribute substantially to the how variable importance measures can be used in
regression effect; thus, the extensive computations and theory-based research. Theoretical considerations can
comparisons that dominance analysis requires may not be factored into both assessments of variable
be justified in such cases. importance and inclusion of variables in regression
models. According to Schafer (1991), the purpose of
Azen and Budescu (2003) explained how theory-based hierarchical regression analysis is to enter
suppression effects can be determined through variables into regression equations in a predetermined
examination of conditional dominance statistics. order that is relevant to the theory underlying
Unique variance contributions of a regular development of the regression model. This process
independent variable averaged over all models of a enables determinations of both the (a) incremental
given size will typically decrease with increasing predictability at each regression step and (b) the
numbers of independent variables. This trend occurs variance explained by the variable(s) entered at each
because variance is typically shared between the step. The researcher is able to control the variance
independent variables, which reduces the amount of contributions of several control variables before
variance an independent variable uniquely explains in entering the variables of primary importance to the
the presence of other independent variables. However, theory with such methods. Researchers can examine
in the suppression case, a suppressor will contribute variable importance in the same hierarchical manner,
more variance on average across subsets with greater and they can frame variable importance in the context
numbers of predictors, as it only contributes variance of the theory underlying the regression model.
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 11
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Table 1. The Multiple Lenses of Multiple Regression: Ways to Operationalize Independent Variable
Contributions to R2
Values are
Always Identical Identifies
Identifies Direct Total Partial
Total When Multi-
Suppressor Effect Effect Effect
R Predictors collinearity
Uncorrelated
Weights X X
Zero-order r X X
Product Measure X X X
Structure
X
Coefficient

Commonality
X X
Coefficients
Unique X X
Common X X
Dominance
Analysis
Complete X X X
Conditional X
General X
Relative Weights X X

Generally, in the presence of strong theory, produce specific effects on a particular dependent
regression equations can be used to answer three variable that are used to inform intervention-related
types of theoretical research questions: (a) can theory in such a case.
specific combinations of independent variables
Stepwise Regression Methods
predict or explain variance in the dependent
variable?; (b) is a specific variable in a set of Stepwise regression methods are sometimes
independent variables necessary to predict or relied upon to determine a set of independent
explain variance in the dependent variable; and (c) variables that purportedly represent the best set of
can specific combinations of independent variables predictors of a particular dependent variable.
predict or explain variance in the dependent Hinkle, Wiersma, and Jurs (2003) outlined the steps
variable, given a strong theoretical rationale for involved in conducting forward stepwise regression.
including control variables as predictors? (Thayer, The first independent variable is selected for entry
2002). In the context of intervention research, into the regression equation that demonstrates the
Groemping (2007) also argued that researchers can highest bivariate correlation with the dependent
use regression results to make comparative variable. The second independent variable selected
judgments of which independent predictors can produces the highest increase in R2 after accounting
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 12
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

for the prediction of the first variable. After this fewer independent variables may predict an
second independent variable is added, a second equivalent R2 as the models selected by the stepwise
significance test is conducted to determine if the solution; (c) independent variables not included in
first independent variable remains a statistically the stepwise solution may be just as significant
significant predictor; if it is not, it is dropped from independent variables as those that are included; and
the equation. This process repeats until either (a) all (d) the independent variables will not enter the
independent variables have been entered into the model in the order of importance that they would in
equation or (b) entry of the remaining independent a final model of independent variables selected
variables into the stepwise solution does not simultaneously (Thayer, 1990).
produce a statistically significant increase in R2. Lastly, and perhaps most importantly, selection
There are several significant difficulties inherent of variables in stepwise solutions capitalizes hugely
in stepwise regression analyses (Thompson, 1989, on sampling error (Thompson, 1995). An
1995) that caution against its widespread use as a independent variable that contributes an amount of
measure of variable importance. First, as variance that explains an extremely miniscule
Thompson (1995) pointed out, the analysis does not amount of greater variance than another variable
incorporate the correct number of degrees of due to sampling error may be chosen as the best
freedom for each statistical significance test. The predictor for a particular step, and all following
method assigns one degree of freedom to each variable selections will thereby result from a
variable that is tested at each step; however, because chance selection at this previous step. Due to
each independent variable is selected from all these three major problems with the stepwise
remaining independent variables in the regression regression method, we generally proscribe use of
equation at each step, a degree of freedom should this technique in assessing variable importance. We
be assigned to all considered variables. This dedicated considerable discussion to these issues
oversight increases the F value at each step and its because the method remains in popular use by
likelihood of being statistically significant. The researchers.
miscalculation of degrees of freedom will thus result
Illustration
in more Type I errors overall (rejecting the null
hypothesis when it is true; Thompson, 1989). Below we present an example of a results
section that employs all variable importance
Even if this issue were addressed, a second
measures discussed in our guidebook to interpreting
significant difficulty with the stepwise regression
MR results, in hopes that this can serve as a
procedure is that the independent variables that are
template for other researchers to use in their own
selected are conditional on the variance
work. Data for this example was obtained from the
contributions of the variables that have already been
Holzinger and Swineford (1939)s dataset. This
entered into the regression (Thompson, 1995). All
dataset contains assessments of 301 subjects from
results hinge on the variable that is selected as the
two high schools (Paster School and White Grant
first predictor; accordingly, the researcher will get
School) on a battery of 26 tests that measured
different entry orders at different steps depending
verbal, spatial, and mathematical abilities. This
on the variable that starts the stepwise solution.
example used three tests relevant to subjects
Because of this conditionality, the stepwise
mathematical aptitude: (a) numeric, (b) arithmetic, and
regression process does not answer the question
(c) addition to predict deductive mathematical reasoning
that it claims to answer, that is, what is the best set
(i.e., reasoning).
of independent variables to predict variance in a
particular dependent variable? In fact, it is likely if The regression equation for this analysis was:
stepwise regression methods are used that (a) other reasoning = -2.92 + 1.43numeric + 0.89arithmetic
models with the same number of independent 0.12addition. A comparison across all statistics
variables may have a larger R2; (b) models with presented in Table 2 highlighted that numeric was the
strongest direct predictor of reasoning across multiple
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 13
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Table 2. Summary of Statistics Determining Independent Variable Contributions to Regression Effects


Variable rs rs2 r Pratt Unique Common GDW RWI RWI %
Numeric .35 .87 .76 .40 .140 .092 .065 .124 .124 60.0
Arithmetic .23 .72 .52 .33 .075 .038 .070 .072 .073 35.5
Addition -.16 .09 .01 .04 -.006 .022 -.020 .011 .009 4.5

indices. Numeric obtained the largest beta weight ( explained a large portion of the overall regression
= .35, p < .001), demonstrating that it made the effect (.124, 60%) when partitioning that effect
largest contribution to the regression equation, while based on creation of variables uncorrelated or
holding all other predictor variables constant. The independent counterparts. Dominance analysis
zero-order correlation of numeric with the reasoning (r results (see Table 4) demonstrated complete
= .40), when squared, showed that numeric shared dominance of numeric over arithmetic and addition, as it
the largest amount (16%) of its variance with contributed more unique variance in the regression
reasoning. The squared structure coefficient (rs2 = .76) effect than the other two independent variables
demonstrated that numeric explained the largest across all 3 MR sub-models that include that
amount (76%) of the variance in , the predicted variable. This complete dominance can be
values of reasoning. Product measure results determined by looking across each row in Table 4
demonstrated that numeric accounted for the largest and seeing how each value for numeric was larger
partition of variance in reasoning (.140, 67.6% of the than the values for arithmetic and addition.
Table 3. Commonality Coefficients Table 4. Complete Dominance Weights
Coeffic Additional
Effect Percent Contribution of:
-ient
Unique to Numeric .092 44.5 Model Num- Arith- Addi-
Variable(s)
R eric metic tion
Unique to Arithmetic .038 18.5
Subset Containing
Unique to Addition .022 10.7 .158 .108 .002
No Predictors
Common to Numeric and .075 36.2 Numeric .152 .027 .011
Arithmetic
Arithmetic .108 .073 .008
Common to Numeric and -.015* -7.4*
Addition Addition .002 .167 .113
Common to Arithmetic and -.011* -5.3* Numeric and
.185 .022
Addition Arithmetic
Common to Numeric, Arithmetic, .006 2.9 Numeric and
.169 .038
and Addition Addition
Total .207 100.0 Arithmetic and
.115 .092
Addition
*Negative values represent suppression effects
Numeric,
regression effect) when multiplying the beta weight Arithmetic, and .207
(.35) by the zero-order correlation (.40). Notably, Addition
relative weight results supported that numeric
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 14
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Because numeric was a completely dominant demonstrating that arithmetic accounted for 35.5% of
independent variable over arithmetic and addition, it the regression effect when partitioning it based on
necessarily showed conditional dominance (see creation of the independent variables uncorrelated
Table 5) over arithmetic and addition. This conditional counterparts. Thus, arithmetic accounted for the
dominance can be determined by looking across all second largest amount of variance in the regression
rows in Table 5 and seeing how numeric contributed equation across multiple measures.
more unique variance on average to regression Complete dominance analysis results (see Table
effects for models of all subset sizes than arithmetic 4) supported the fact that arithmetic was completely
and addition. As numeric was completely and dominant over addition (see Table 4), as a perusal of
conditionally dominant over the other two all rows in Table 4 show that it contributed more
predictors in the model, it also showed general unique variance to all rows of all subsets sizes than
dominance over both of these variables (see Table addition. As arithmetic was completely dominant over
2). These statistics are equal to the averages of the addition, it also showed conditional dominance (see
conditional dominance weights shown in Table 5. Table 5) in and general dominance (see Table 2)
Notably, for this example, the relative weight (.124) over this variable. Once again, the general
was equivalent to the general dominance weight dominance weight for arithmetic was nearly identical
(.124); thus, its contribution to the regression effect to the relative weight, reflecting that arithmetic
assessed in terms of averages of unique variance explained the second largest and a substantial 35.5%
contributions to all possible subsets or through the of the variance in R when partitioning the
creation of uncorrelated counterpart variables was regression effect based on the average difference in
the same. fit between subsets that do and do not include
Table 5. Conditional Dominance Weights arithmetica different partitioning method than that
used for relative weights.
Subset
Numeric Arithmetic Addition Although other statistics clearly showed how
Size
0 .158 .108 .002 numeric was the strongest direct contributor to the
regression equation, followed by arithmetic, they did
1 .122 .070 .009 not show exactly how those variables contributed
2 .092 .038 .022 unique and shared variance to the regression
equation. Thus, commonality coefficients were
Arithmetic clearly emerged as the second consulted to obtain this information (see Table 3).
strongest direct predictor of reasoning. In terms of When viewing commonality analysis results, unique
the beta weight ( = .23, p < .001), it made the effect results demonstrated that numeric contributed
second largest contribution to the regression more unique variance (44.5%) to the regression
equation when holding all other predictors constant. effect than did arithmetic (18.5%). The common
Its zero-order correlation (r = .33) was also the effect for numeric and arithmetic reflected that both
second largest in the model, which, when squared, variables also contributed substantial shared
demonstrated that arithmetic shared the second variance (36.2%) to the regression effect. These
largest amount (10.9%) of its variance with reasoning. results highlighted that numeric and arithmetic partially
The squared structure coefficient (rs2 = .52) operated in combination in predicting reasoning.
illustrated that arithmetic shared the second largest Viewing unique and common effect columns in
amount (52%) of variance with . Product measure Table 2 illustrated how numeric contributed more
(.075) results showed that arithmetic accounted for unique than shared variance to variance in reasoning
36.2% of R2 when it was partitioned based on (.09 vs. .07, respectively), while, conversely, arithmetic
multiplying the beta weight (.23) by the zero-order contributed more shared than unique variance to
correlation (.33). Arithmetics relative weight (.073) variance in the dependent variable (.04 vs. .07,
was nearly identical to its product measure, respectively).
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 15
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

The results for different measures presented either numeric or arithmetic, suggesting that addition
different lenses on the predictive power of addition to might be a suppressor for both numeric and arithmetic.
predicting variance in the dependent variable. The Commonality coefficient findings supported
beta weight for addition ( = -.16, p = .004) that addition was a suppressor for both numeric and
suggested that addition played a more minor but still
arithmetic, as the common effects between numeric
statistically significant role in the regression effect and addition (-.015) and arithmetic and addition (-.011)
when holding all other variables constant. However, were negative, and suppression is generally
the squared structure coefficient (rs2 = .001) showed demonstrated in negative commonalities. Summing
that addition contributed negligible variance to , the percentages of variance that each of these
and the zero-order correlation (r = .04) showed that common effects contributes to the overall
addition explained 0.2% of the variance in the overall regression effect (7.4% for numeric and 5.3%, for
dependent variable. These statistics clearly addition, respectively) showed that 12.7% of the
demonstrated that addition did not directly share regression effect is due to suppression, and thus that
variance with either the obtained effect or the addition removes 12.7% of irrelevant variance from
dependent variable as a whole. When partitioning numeric and arithmetic to increase the amount of
R based on multiplying the beta weight (-.16) by the variance in the dependent variable explained by
zero-order correlation (.02), product measure (-.006) these two independent variables by 12.7%.
results reflected that addition once again contributed
very little variance to the regression effect (-2.9%). Overall, these findings supported how both
The relative weight for addition (.009) demonstrated numeric was the most significant direct contributor
that addition contributed very little variance the and arithmetic was the second most important direct
regression effect (4.5%) when partitioning it based contributor to predicting variance in reasoning, as
on variables uncorrelated counterparts. The fact reflected across different measures of direct, total, and
that the beta weight for addition was statistically partial effects. It is important to note that this is not
significant but that addition shared little variance with always the case: One independent variable may be
reasoning when examining (a) its shared variance with deemed the most important through one lens, and
, (b) its shared variance with the dependent another independent variable may achieve that
variable, and (c) partitioning the regression effect status through another lens. Results also supported
based on two measures suggested that addition was a from multiple lenses how addition functioned as a
suppressor variable, which removed irrelevant suppressor in this regression equation. Reliance on
variance in at least one independent variable, beta weights alone would not have pointed out the
thereby allowing its (or their) contributions to the nature of the suppressor effect.
regression effect to become larger. Table 6 is offered as a summary of the purposes
Conditional dominance weights confirmed that of each of the discussed statistics.
addition was a suppressor, as its contribution to List of Recommendations for Practice
regression models increased across subset sizes. Its
contribution to the subset containing no other We present a list of recommendations for
predictors was .002, which illustrated that it reporting of regression results based on the methods
contributed nearly negligible variance to the discussed in this guidebook to assist researchers in
regression effect in isolation from other variables in selection and synthesis of multiple variable
the regression equation. Its average unique variance importance measures. Prior to outlining our
contribution for the models containing only one of recommendations, we would like to state a general
the other two predictors was .009, and for the recommendation to run scatterplots of all
subset containing both other predictors it was .022. independent variables and the dependent variable to
Notably, additions contribution to the regression find outliers and/or unexpected patterns in the data
model containing both numeric and arithmetic was prior to conducting MR analyses. Clean up the data
greater than its contribution to models containing
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 16
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Table 6. Purposes of Each Statistical Measure as appropriate before proceeding with summarizing
them with regression and other statistical analyses.
Statistic Purposes of Measure
Also, previous articles have recommended including
Beta weight Determination of independent tables of means, standard deviations, and
variables contributions to prediction intercorrelations between independent variables to
within a linear regression equation accompany regression summary tables (cf. Schafer,
while holding all other independent 1991). Reporting reliability coefficients for all
variables constant variables can also be helpful (e.g., independent
Zero-order Determination of magnitude and variables with low alpha reliability might explain low
r direction of bivariate linear relationship intercorrelations and weak prediction). Previous
between each independent variable and authors have also recommended (a) including the
the dependent variable specific regression equation(s) in a footnote or in
Product Determination of variable importance text (i.e., do not merely list the total R2 or the
measure based on partitioning model R2 as a statistical significance of the equation; cf. Schafer,
function of a predictors beta weight 1992) as well as (b) including a general regression
multiplied by its zero-order r summary table that reports the values of the zero-
Squared Determination of how much variance order correlation between each independent variable
structure each independent variable contributes and the dependent variable, F and MSresidual values,
coefficients to and p values for each individual predictor to
illustrate statistical significance (cf. Schafer, 1992).
Commonality coefficients:
Specific to reporting multiple indices of variable
Unique Determination of variance each importance, we offer following recommendations
effects independent variable contributes to a for practice:
regression equation that is not shared
with other independent variables 1) Do not rely solely upon beta weights when
(squared semipartial correlation) interpreting regression findings, except in
the case of uncorrelated predictors or when
Common Determination of which independent
effects variables share variance in predicting the model is perfectly specified (see, e.g.,
the dependent variable as well as Courville & Thompson, 2001). Beta weights
quantification of how much variance is are only informative with regard to
shared between independent variables prediction; they do not tell the researcher
other important information provided by
Dominance Determination of whether one
weights independent variable contributes more the other metrics we have reviewed.
variance than another independent 2) Use different tables to help interpret
variable to models: different indices of variable importance.
a. Include one table that enables visual
(a) across all subsets of
comparisons across indices for each
independent variables
(complete dominance)
independent variable. Joint
(b) on average across models for comparisons across indices can aid
all subset sizes (conditional in identification of associations
dominance) between variables and the presence
(c) on average across all models of suppression in a regression
(general dominance) equation. Table 2 of this paper is
Relative Determination of variable importance recommended if all indices are
weights based on method that addresses compared. If specific subsets of
multicollinearity by creating variables indices are used for a specific
uncorrelated counterparts research purpose, then one can still
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 17
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

present those indices in the format structure coefficient results (supportive of


we display. multicollinearity/associations between
b. If commonality analysis is used, variables and/or suppression), it is suggested
include a table (see Table 3) that lists that researchers employ commonality
the unique variance contributions of analysis to determine the location and extent
each independent variable and the of this shared variance and/or suppression.
common variance contributions for 6) Include a variance partitioning statistic in
all possible subsets of independent addition to beta weights in the presence of
variables to the regression equation. correlated predictors (e.g., general
Also report these portions of R2 as dominance weights, relative weights, and
percentages of R2, as well. Pratts measure). Draw comparisons in text
c. If dominance analysis is used, between techniques that partition R2 if
include a table (see Table 4) of multiple techniques are used.
unique variance contributions of 7) In the presence of suppression in a
each independent variable to all regression equation, always calculate
possible subset sizes to enable commonality coefficients, as this technique
pairwise comparisons of unique is uniquely able to identify which variables
variance contributions of each are being suppressed and quantify the
independent variable to all subsets of overall amount of suppression present in an
predictors. Additionally, include a equation.
table (see Table 5) of conditional 8) Always consider the statistics used in
dominance weights that averages calculating each index when evaluating
unique variance contributions of variable importance. For example, if a Pratt
each independent variable to models measure value is near-zero, the researcher
of all subset sizes. This table is useful should verify that the zero-order correlation
in identifying suppression effects if used in computing the Pratt measure and
values increase across subset sizes not the beta weight shows a near-zero or
with more independent variables. zero value when determining if a variable is
3) When reporting commonality analysis results a suppressor, as suppression is only
in text, describe both the unique and shared demonstrated in the former case. The
variance contributions of all independent researcher should also consider the value of
variables and whether each variable R2 that is expressed as percentages in
contributes more shared or unique variance commonality analysis and relative weights
in its contribution to R2. analysis. It is reasonable to think that
4) When reporting dominance analysis results percentages of a small R2 might be
in text, include all three types of dominance interpreted differently than the same
(complete, conditional, and general), as percentages based on a large R2.
dominance relationships can be established
Conclusion
at lesser levels of dominance (i.e., general
dominance) if not at higher levels (i.e., This paper has illustrated how researchers
complete dominance). conceptualizations and assessments of variable
5) Calculate squared structure coefficients importance can be enhanced by viewing MR results
when independent variables are correlated in through multiple lenses. As it has shown that each
order to determine the role of shared lens has distinct advantages and limitations, and
variance in the regression equation (cf. that multiple statistical measures complement each
Courville & Thompson, 2001). If there is other in the perspectives they provide regarding
divergence between beta weight and regression findings, we hope that this paper will
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 18
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

encourage researchers to look beyond beta meeting of the New England Educational Research
weights and employ the other measures discussed in Association, Provincetown, MA.
our guidebook. Ideally, this practice would become Fabbris, L. (1980). Measures of predictor variable
a matter of routine among researchers, in our peer- importance in multiple regression: An additional
reviewed journals, and in teaching MR within suggestion. Quality & Quantity, 4, 787792.
graduate-level statistics curricula. The hope is that Genizi, A. (1993). Decomposition of R2 in multiple
regression with correlated regressors. Statistica
our data-driven example will allow researchers to
Sinica, 3, 407-420.
write up their own findings in a similar manner and
Grmping, U. (2007). Estimators of relative importance
thus will be able to better represent the richness of in linear regression based on variance
their regression findings, and how these findings are decomposition. The American Statistician, 61, 139-
impacted by such issues as suppression and patterns 147.
of shared variance that go undiscovered through Hinkle, D. E., Wiersma, W., & Jurs, S. G. (2003). Applied
heavy reliance on beta weights alone. statistics for the behavioral sciences (5th ed.). Boston,
MA: Houghton Mifflin.
References
Holzinger, K. J., & Swineford, F. (1939). A study in factor
Amado, A. J. (1999, January). Partitioning predicted variance analysis: The stability of a bi-factor solution (No. 48).
into constituent parts: A primer on regression commonality Chicago, IL: University of Chicago.
analysis. Paper presented at the annual meeting of Hoyt, W. T., Leierer, S., & Millington, M. (2006).
the Southwest Educational Research Association, Analysis and interpretation of findings using
San Antonio, TX. multiple regression techniques. Rehabilitation
Azen, R., & Budescu, D. V. (2003). The dominance Counseling Bulletin, 49, 223-233.
analysis approach for comparing predictors in Johnson, J. W. (2000). A heuristic method for estimating
multiple regression. Psychological Methods, 8(2), 129- the relative weight of predictor variables in multiple
148. doi: 10.1037/1082-989X.8.2.129 regression. Multivariate Behavioral Research, 35(1), 1-
Baltes, B. B., Parker, C. P., Young, L. M., Huff, J. W., & 19.
Altmann, R. (2004). The practical utility of Johnson, J. W., & LeBreton, J. M. (2004). History and
importance measures in assessing the relative use of relative importance indices in organizational
importance of work-related perceptions and research. Organizational Research Methods, 7(3), 238-
organizational characteristics on work-related 257. doi: 10.1177/1094428104266510
outcomes. Organizational Research Methods, 7(3), 326- LeBreton, J. M., Ployhart, R. E., & Ladd, R. T. (2004). A
340. doi: 10.1177/1094428104266014 Monte Carlo comparison of relative importance
Budescu, D. V. (1993). Dominance analysis: A new methodologies. Organizational Research Methods, 7(3),
approach to the problem of relative importance of 258-282. doi: 10.1177/1094428104266017
predictors in multiple regression. Psychological Mayeske, G. W., Cohen, W. M., Wisler, C. E., Okada, T.,
Bulletin, 114(3), 542-551. Beaton, A. E., Proshek, J. M., Weinfeld, F. D., &
Capraro, R. M. & Capraro, M. M. (2001). Commonality Tabler, K. A. (1969). A study of our nations schools.
analysis: Understanding variance contributions to Washington, DC: U.S. Department of Health,
overall canonical correlation effects of attitude Education, and Welfare, Office of Education.
toward mathematics on geometry achievement. Mood, A. M. (1969). Macro-analysis of the American
Multiple Linear Regression Viewpoints, 27(2) 16-23. educational system. Operations Research, 17, 770-784.
Courville, T., & Thompson, B. (2001). Use of structure Mood, A. M. (1971). Partitioning variance in multiple
coefficients in published multiple regression articles: regression analyses as a tool for developing learning
is not enough. Educational and Psychological models. American Educational Research Journal, 8, 191-
Measurement, 61, 229-248. 202.
Darlington, R. B. (1968). Multiple regression in Newton, R. G., & Spurrell, D. J. (1967). A development
psychological research and practice. Psychological of multiple regression for the analysis of routine
Bulletin, 69(3), 161-182. data. Applied Statistics, 16, 51-64.
DeVito, P. J. (1976, May). The use of commonality analysis in Nimon, K. (2010). Regression commonality analysis:
educational research. Paper presented at the annual Demonstration of an SPSS solution. Multiple Linear
Regression Viewpoints, 36(1), 10-17.
Practical Assessment, Research & Evaluation, Vol 17, No 9 Page 19
Nathans, Oswald & Nimon, Interpreting Multiple Regression Results

Nimon, K., Gavrilova, M., & Roberts, J. K. (2010). Thayer, J. D. (1990, April). Implementing variable selection
Regression results in human resource development techniques in regression. Paper presented at the annual
research: Are we reporting enough? In C. Graham meetings of the American Educational Research
& K. Dirani (Eds.), Proceedings of the Human Resource Association, Boston, MA.
Development 2010 International Conference (pp. 803- Thayer, J. D. (2002, April). Stepwise regression as an
812), Knoxville, TN: AHRD. exploratory data analysis procedure. Paper presented at
Nimon, K., Lewis, M., Kane, R., & Haynes, R. M. (2008). the annual meetings of the American Educational
An R package to compute commonality coefficients Research Association, New Orleans, LA.
in the multiple regression case: An introduction to Thomas, D. R., Hughes, E., & Zumbo, B. D. (1998).
the package and a practical example. Behavior On variable importance in linear regression. Social
Research Methods, 40, 457-466. doi: Indicators Research: An International and Interdisciplinary
10.3758/BRM.40.2.457 Journal for Quality-of-Life Measurement, 45, 253-275.
Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric Thompson, B. (1989). Why wont stepwise methods die?.
theory (3rd ed.). New York: McGraw-Hill, Inc. Measurement and Evaluation in Counseling and
Onwuegbuzie, A. J., & Daniel, L. G. (2003, February 19). Development, 21, 146-148.
Typology of analytical and interpretational errors in Thompson, B. (1995). Stepwise regression and stepwise
quantitative and qualitative educational research. discriminant analysis need not apply here: A
Current Issues in Education [On-line], 6(2). Available: guidelines editorial. Educational and Psychological
http://cie.ed.asu.edu/volume6/number2/ Measurement, 55(4), 525-534.
Pedhazur, E. J. (1997). Multiple regression in behavioral Tonidandel, S., & LeBreton, J. M. (2010). Determining
research: Explanation and prediction (3rd ed.). Stamford, the relative importance of predictors in logistic
CT: Thompson Learning. regression: An extension of relative weight analysis.
Pratt, J. W. (1987). Dividing the indivisible: Using simple Organizational Research Methods, 13(4), 767-781. doi:
symmetry to partition variance explained. In T. 10.1177/10944281109341993
Pukkila & S. Puntanen (Eds.), Proceedings of the second Tonidandel, S., & LeBreton, J. M. (2011). Relative
international Tampere conference in statistics (pp. 245- importance analysis A useful supplement to
260). Tampere, Finland: University of Tampere. regression analyses. Journal of Business and Psychology,
Rowell, R. K. (1996). Partitioning predicted variance 26, 1-9. doi: 10.1007/s10869-010-9204-3
into constituent parts: How to conduct regression Zientek, L. R., Capraro, M. M., & Capraro, R. M. (2008).
commonality analysis. In B. Thompson (Ed.), Reporting practices in quantitative teacher
Advances in social science methodology (Vol. 4, pp. 33-43). education research: One look at the evidence cited
Greenwich, CT: JAI Press. in the AERA panel report. Educational Researcher, 34,
Schafer, W. D. (1991). Reporting hierarchical regression 208-216. doi: 10.3102/0013189X08319762
results. Measurement and Evaluation in Counseling and Zientek, L. R., & Thompson, B. (2006). Commonality
Development, 24, 98-100. analysis: Partitioning variance to facilitate better
Schafer, W. D. (1992). Reporting nonhierarchical understanding of data. Journal of Early Intervention,
regression results. Measurement and Evaluation in 28, 299-307. doi: 10.1177/105381510602800405
Counseling and Development, 24, 146-149. .
Citation:
Nathans, Laura L., Oswald, Frederick L., & Nimon, Kim (2012). Interpreting Multiple Linear Regression: A
Guidebook of Variable Importance. Practical Assessment, Research & Evaluation, 17(9). Available online:
http://pareonline.net/getvn.asp?v=17&n=9

Corresponding Author:

Laura L. Nathans
University of North Texas
laura_nathans [AT] yahoo.com

You might also like