You are on page 1of 20

Industrial and Organizational Psychology, 4 (2011), 3–22.

Copyright © 2011 Society for Industrial and Organizational Psychology. 1754-9426/11

FOCAL ARTICLE

Evidence-Based I–O Psychology:


Not There Yet

ROB B. BRINER
University of London

DENISE M. ROUSSEAU
Carnegie Mellon University
Abstract
Evidence-based practice is now well established in several fields including medicine, nursing, and social policy.
This article seeks to promote discussion of whether the practice of industrial–organizational (I–O) psychologists
is evidence based and what is needed to make I–O psychology an evidence-based discipline. It first reviews
the emergence of the concept of evidence-based practice. Second, it considers the definitions and features of
evidence-based practice, including evidence-based management. It then assesses whether I–O psychology is
itself an evidence-based discipline by identifying key characteristics of evidence-based practice and judging the
extent these characterize I–O psychology. Fourth, some key strategies for promoting the use of evidence in
I–O psychology are considered: practice-oriented research and systematic reviews. Fifth, barriers to practicing
evidence-based I–O psychology are identified along with suggestions for overcoming them. Last is a look to the
future of an evidence-based I–O psychology that plays an important role in helping consultants, in-house I–O
psychologists, managers, and organizations become more evidence based.

Most industrial–organizational (I–O) psy- in journal articles or workforce facts and


chologists would confidently claim that metrics collected through practice.
their practice is based on evidence—in Look a bit deeper, however, and things
particular, psychological research findings. aren’t quite so straightforward. Practitioners
And, in a general sense, that’s more or less of all stripes—from the snake-oil salesper-
true. The founders of our field, including son with a cure-all remedy to the fam-
James McKeen Cattell and C. S. Myers, ily physician or the personal trainer to
originated the application of systematic the psychic who communicates with the
research to workplace issues and in doing dead—claim to have evidence for what
so established psychological science itself they do. Claims to use evidence are there-
in both the United States and Britain. Most fore meaningless in themselves. However,
of what we do in I–O psychology draws on the hallmark of any profession is the exis-
or is at least informed by some type of evi- tence of an agreed-upon core of knowledge
dence whether that be research published and means for its continued generation and
refinement (Friedson, 1986). Unlike com-
mon sense or general knowledge claims,
Correspondence concerning this article should be it’s the nature and quality of the par-
addressed to Rob B. Briner.
E-mail: r.briner@bbk.ac.uk ticular evidence a profession’s practition-
Address: Department of Organizational Psychol- ers use that distinguishes them from the
ogy, Birkbeck College, University of London, Malet layperson—or the snake-oil salesperson. In
Street, London WC1E 7HX, United Kingdom.
We are very grateful to three anonymous reviewers the case of professions such as I–O psychol-
for their comments and suggestions. ogy, medicine, education, or engineering,
3
4 R.B. Briner and D.M. Rousseau

much of their value and legitimacy depends majority of respondents (who were largely
on the extent a scientific evidence base practitioners) held the opinion that prac-
informs and is used in practice. tice is ahead of research in 14 content
In many areas of I–O psychology, large areas, including coaching, talent manage-
bodies of reasonable quality evidence are ment, and employee relations, whereas
continually refined and critically evaluated. science was ahead in only two (measure-
The expanded use of meta-analysis in I–O ment and statistics, and job analysis). In five
psychology indicates the value it places on areas, practitioners and nonpractitioners
systematic evidence, cumulative research held opposite views. Practitioners saw prac-
findings, and critical analyses (e.g., Judge, tice ahead and nonpractitioners saw science
Heller, & Mount, 2002; Judge & Illies, ahead in recruitment, performance manage-
2002). Similarly, the value of the structured ment, organizational culture, training and
interview over its unstructured counterpart development, and employee engagement.
has been affirmed repeatedly over decades A plurality shared the view that science is
(McDaniel, Whetzel, Schmidt, & Maurer, behind practice in its choice of research
1994); the I–O psychology practitioner topics and that practice-relevant research
is probably one of the more consistent was sparse. Finding perceived gaps on both
voices in industry to buck the populist sides is not so surprising. In I–O psy-
tide that continues to favor unstructured chology scientists and practitioners each
techniques. On the other hand, I–O prize their own knowledge sources over
psychology does not always resist more the other’s, raising concern regarding the
‘‘pop-psychology’’ topics such as emotional quality of the interface between the two. In
intelligence. Although introduced to social turn, these findings raise questions about
science through work by Mayer and the extent the science and practice of
Salovey (1997), the type of claims made I–O psychology is synergistic in a fash-
in Daniel Goleman’s (1998) popular book ion that would promote evidence-based
and other popular accounts of the topic are practice.
clearly questionable (e.g., Cherniss, 2010; Another expression of this concern can
Locke, 2005). be found in many of this journal’s focal
In other areas, the evidence familiar to articles, as they are precisely about exam-
I–O psychologists may appear more static, ining the concepts and evidence under-
taking the form of received wisdom. For lying our practice in a range of areas
example, consider the notion that accurate including the identification of discrimi-
feedback is good (i.e., at least if task nation in workplace evaluations (Landy,
related; cf., Kluger & DeNisi, 1996). Still, 2008), executive coaching (McKenna &
uncritical reliance on received wisdom Davis, 2009), job performance ratings
is problematic: Cognitive and educational (Murphy, 2008), employee engagement
research demonstrates that feedback too (Macey & Schneider, 2008), executive
early in the learning process can undermine selection (Hollenbeck, 2009), and leader-
the mental processes needed to acquire ship development (McCall, 2010). These
certain skills (Merrill, Reiser, Merrill, & articles have created lively, important, and
Landes, 1995; VanLehn, Siler, Murray, healthy debate.
Yamauchi, & Baggett, 2003). Given that Many professions have expressed such
new information replaces or refines existing concerns and pursued ways that evidence
information, the evidence base we use is can better inform practice. One of the
inevitably a work in progress. most recent and widespread ideas used
I–O psychologists have expressed con- to frame solutions to this problem is the
cerns about the existence, relevance, and subject of this focal article: evidence-based
use of evidence in the profession. A recent practice. For some years, evidence-based
survey of SIOP members (Silzer, Cober, practice has evolved as a way of identifying
Erickson, & Robinson, 2008) revealed that a and developing techniques and processes
Evidence-based I–O psychology 5

that practitioners can use to incorporate medicine and its allied disciplines, such
evidence from various sources into their as nursing, have engaged the challenge of
everyday work. Our purpose here is to pro- using scientific evidence to better inform
mote discussion and further development practice.
of the practice of I–O psychology as an In medicine and nursing, the notion of
evidence-based discipline. being evidence based is now well estab-
This article is structured as follows. lished. It forms part of the basic training
It reviews the emergence of evidence- of nurses, physicians, and other profes-
based practice in medicine, social pol- sions allied to medicine. Medical research
icy, and more recently management, and and its funders, in addition to producing
considers the definitions and features of new basic science, also put considerable
evidence-based practice, including that of resources into research on effective prac-
evidence-based management (EBMgt). It tice, including how to best treat specific
assesses whether I–O psychology is itself types of patients, as well as in conduct-
an evidence-based discipline by identify- ing systematic reviews of research litera-
ing key characteristics of evidence-based ture to answer practice-relevant questions.
practice and gauging the extent these char- Systematic reviews are essentially litera-
acterize the practice of I–O psychology. By ture reviews that address a very specific
identifying the key contributions of practice- review question using an explicit, system-
oriented research and systematic reviews to atic methodology to identify, select, and
evidence-based practice, this article details pull together findings from a range of studies
how these applications promote evidence to draw conclusions about what is known
use in practice, and describes barriers to and not known about the question. (Meta-
evidence-based approaches to I–O psy- analyses are one type of systematic review.)
chology practice along with suggestions to Systematic reviews are in essence pieces of
overcome them. Last, we look to the future research on existing research.
and consider the prospects for an evidence- The findings of both patient-oriented
based I–O psychology that helps managers studies and systematic reviews are then
and organizations themselves become more translated into tools and decision-making
evidence based. aids such as checklists and patient-care
protocols used by medical clinicians. Later,
Emergence of Evidence-Based we will talk about how the ‘‘functional
Practice and Management I–O psychology equivalents’’ of these can
The notion of using scientific evidence to promote our own evidence-based practice.
inform professional practice is not new. At In the late 1990s, the term ‘‘evidence
the same time, neither is the observation based’’ became paired with other nonmedi-
that from medicine to management much cal disciplines and practice areas, including
practice is not related to evidence in education, social work, criminology, and
any significant way. This tension found government policy making. Over a decade
expression in a British Medical Journal ago, Briner (1998) made what appears to
editorial reporting, ‘‘only about 15% of be the first attempt to suggest that some
medical interventions are supported by ideas from evidence-based medicine could
solid scientific evidence’’ (Smith, 1991, be adapted to the practice of organizational
p. 798).1 This article marked a turning psychology. These same ideas have been
point in the development of evidence-based applied more recently to the practice of
medicine. More than other professions, management. In 2006, two key publications
(Pfeffer & Sutton, 2006; Rousseau, 2006)
stimulated discussions of EBMgt and how it
1. It may be possible and potentially useful to might be used in business school teaching
make a similar estimate of the percentage of
I–O psychology practices that are based on solid (Rousseau & McCarthy, 2007), its limita-
scientific evidence. tions and potential dangers (Learmonth &
6 R.B. Briner and D.M. Rousseau

Harding, 2006; Morrell, 2008), the evi- What Is Evidence-Based Practice?


dence for EBMgt (Reay, Berta, & Kohn,
As mentioned, all practitioners claim
2009), and its meaning and definition
to have evidence for their practice.
(Briner, Denyer, & Rousseau, 2009).
Nonetheless, evidence-based practice is a
In the last 2 decades, considerable soul
particular approach or more accurately
searching by management scholars over
a set of approaches to incorporate evi-
the research–practice ‘‘gap’’ or ‘‘divide’’
dence into practice decisions. In medicine,
has raised difficult questions about why for example, this means ‘‘integrating indi-
the gap exists, how to bridge it, and the vidual clinical expertise with the best
value and purpose of management research available external clinical evidence from
itself (Hambrick, 1994; Rynes, Brown, & systematic research’’ (Sackett, Richardson,
Colbert, 2002). Although I–O psychol- Rosenburg, & Haynes, 1997, p. 2) in making
ogy is not immune to such concerns, in decisions about patient care. Three aspects
other respects I–O psychology would seem of this definition need to be highlighted.
to have avoided the criticisms levied on First, evidence-based practice integrates
other social sciences (Anderson, Herriot, & the practitioner’s expertise and external
Hodgkinson, 2001). Anderson and col- evidence from research. Both sources of
leagues used the label ‘‘pedantic science’’ knowledge are vital. Second, it is about
to describe those research domains driven trying to obtain and use the best available
by theoretical concerns and fastidious ana- evidence even if ultimately determined to
lytics while ignoring real-world issues. be inconsistent or rejected as irrelevant.
I–O psychology might seem an unlikely Using evidence does not mean slavishly
candidate for the pedantic label. Even when following it, acting only when there is
conducted by university-based scholars, good evidence or doing nothing if there is
I–O psychology research is rooted in issues none. Rather, it is a more active process of
and problems arising in organizational examining and evaluating the best of what
settings. Given that its graduates work in is there and applying it along with other
industry as consultants or in-house I–O sources of information, such as situational
psychologists, our discipline would seem facts, to decision making. Some research
to be an exemplar of ‘‘evidence-based’’ evidence may be more readily converted
practice. The research-intensive training of into practical knowledge, although both
I–O psychologists and our often close-to- basic scholarly and more applied evidence
practice research should make reliance on can contribute to evidence-based practice.
research evidence almost second nature. Third, it uses systematic reviews to assess
Indeed, it is probably no coincidence that all available and relevant evidence rather
many advocates of EBMgt (including the than relying on single studies.
present authors) are also I–O psychologists EBMgt has been defined in several
(e.g., Wayne Cascio, Edwin Locke, and Ed different ways but most definitions draw
Lawler). on ideas of evidence-based practice found
So how can we explore the truth of I–O in medicine and elsewhere. One recent
psychology’s claim to engage in evidence- definition of EBMgt (Briner et al., 2009,
based practice? On one hand, I–O psy- p. 19) is as follows:
chology seems to embody Anderson et al.’s
(2001) ideal of a ‘‘pragmatic science,’’ Evidence-based management is about
addressing questions of theoretical and making decisions through the consci-
practical relevance in a methodologically entious, explicit, and judicious use of
sound manner. Yet, if we use the nature four sources of information: practitioner
of evidence-based practice as the starting expertise and judgment, evidence from
point in specifying our criteria, we would the local context, a critical evaluation
draw a different conclusion. of the best available research evidence,
Evidence-based I–O psychology 7

and the perspectives of those people who sources, it may be judged to be of poor
might be affected by the decision. quality and therefore not that usable or
relevant. Third, even where the quantity
The conscientious use of the four and quality of information is relatively bal-
sources of information means that an anced across sources, it may be that one
EBMgt approach involves paying careful highly pertinent piece ‘‘trumps’’ others. In
and sustained attention to sources of what a safety-critical organization, for example,
can be potentially different, conflicting, even where nearly all the information sug-
and sometimes difficult-to-interpret infor- gests that a particular intervention may be
mation. Being explicit means using informa- effective, a very small piece of information
tion from each source in a clear, conscious, implying that it also increases errors may
and methodical way such that the roles be enough to push the decision away from
played by all the information in the final what most of the evidence would suggest.
decision are understood. And being judi- Evidence-based practice across various
cious involves using reflective judgment to fields uses similar approaches intended to
evaluate the validity and relevance of the improve the process and consequences of
information from each source. Evidence and decision making by collecting and critically
information is critically evaluated in relation analyzing evidence from a number sources
to the practice context and problem. and then integrating it to make a practical
Take the example of a senior HRM decision or solve a problem (e.g., Gough,
practitioner in a large firm who believes Kiwan, Sutcliffe, Simpson, & Houghton,
the organization has a problem with high 2003; Soukup, 2000).
absence levels and wants to intervene to
reduce the absence rate. Table 1 provides
How Evidence-Based Is
examples of the sorts of information
I–O Psychology?
that the practitioner may compile and
critically evaluate to decide on what kind So, how evidence based are we? To be
of intervention, if any, is likely to be asked this question can feel a bit unnerv-
most effective. After this information and ing or even insulting. I–O psychology is a
evidence is compiled, the next stage is to science after all. We all know how to read
integrate the different sources of evidence. journals and do research. We understand
The decision takes place at the intersection scientific principles and can distinguish
of these four information sources. Exactly good research from bad. We can interpret
how these sources of information are and apply findings from studies. We eval-
integrated depends on the context and who uate, measure, and assess what we do as
is making the decision. It is likely that they we go along. Although all these statements
will vary in several ways. may be broadly true, they don’t reveal how
First, varying quantities of informa- evidence based we actually are as profes-
tion are available from each source. For sionals. Instead, these statements express
example, little may be known about the some of our principles and aspirations. Such
perspectives of those who may be affected statements may be our expectations, but
by the intervention, but the practitioners they can differ from what we actually do.
involved may have much expertise in and We now turn to examine the extent
experience with the problem. Very strong to which characteristics of evidence-based
asymmetries may lead to decisions biased practice identified above are present in
toward those sources of information of the practice of I–O psychology. Note that
which there simply are more. A second no systematic study exists on the actual
way in which these sources of informa- practice of I–O psychologists, that is, the
tion will vary is in relation to their quality, consultants, in-house psychologists, and
validity, and reliability. Although there may others working in industry who hold I–O
be plenty of information from one of these psychology degrees either at the master’s or
8 R.B. Briner and D.M. Rousseau

Table 1. Examples of Information From Each of the Four Sources Relevant to Intervening
in the Presenting Problem of High Absence

Practitioner expertise and judgment Evidence from the local context


• Have I seen this before? • What actually is the absence rate?
• What happened? • What type of absences and where?
• What are my beliefs about the causes of • What are local explanations for
absence? absence?
• What’s worked in the past and why? • Internal research (e.g., surveys)
• What are my hunches? • What absence management is currently
• What do I think are the causes and in place, and is it working?
possible solutions? • What do managers think is going on?
• Is this situation occurring elsewhere? • What are the possible costs and benefits
• How relevant and applicable is my of interventions? Is it worth intervening
experience? here?
• What is happening or what is going to
Critical evaluation of best available research happen in the organization or outside
evidence it that might be affecting absence?
• What are the average rates of absence
in my sector and location—Is the Perspectives of those who may be affected
absence rate here ‘‘high’’? by intervention decision
• What does systematically reviewed • How do employees feel about the
research evidence suggest to be the proposed interventions?
major causes of absence? • Do they see downsides or unintended
• How relevant and applicable is that negative consequences?
evidence here? • How do managers feel about these
• What does research evidence from interventions?
systematic reviews suggest as effective • How practical or workable do those
interventions? responsible for implementing the
• How well might the interventions the interventions feel?
research describes work here? • What alternative explanations and
proposed solutions do others have?

doctoral level. So, our starting point is this to obtain continuing education credit), our
question: If I–O psychology were strongly interpretations of results from the recent
evidence based in the ways defined and surveys of the I–O psychology profession
described above, what would we expect to and practitioner–research connections dis-
observe? cussed briefly above (Silzer et al., 2008),
Table 2 describes some of the charac- and also decades of teaching in I–O psy-
teristics we would expect in an area of chology programs. Our assessments of I–O
practice if it were also evidence-based psychology practice are discussed below,
per the above definitions. It describes our each accompanied by an explanation.
current judgment of the extent these char-
acteristics are observable in I–O psychol- 1. The term ‘‘evidence based’’ is used or
ogy. Others will have different experi- known. Although the general notion
ences and observations, which we hope of using evidence is well established,
will be expressed in the responses to this the specific term ‘‘evidence-based’’
focal article. Our judgments are based on and what it entails is not widely used
observations from a number of sources, or well known. It is rare to find the
including our recent participation in the term ‘‘evidence based’’ paired with
first evidence-based practice workshop at I–O psychology or organizational
SIOP (attended by dozens of practitioners psychology. However, the 2009 SIOP
Evidence-based I–O psychology 9

Table 2. Some Key Characteristics of Evidence-Based Practice and an Estimate of the


Extent to Which They Are Found in I–O Psychology

Found in I–O
Some characteristics of evidence-based practice psychology?
1. The term ‘‘evidence based’’ is used or well known. Given To a very limited
evidence-based practice exists in many other fields and the term is extent?
widely used, it is unlikely that any field adopting an evidence-based
approach would not know of and use the term, even though
definitions might vary across practice fields.
2. The latest research findings and research summaries are accessible. To a very limited
It is difficult to do evidence-based practice without access to evi- extent?
dence in research journals and research summaries. A fundamental
principle of evidence-based practice is that systematic reviews of the
best available external evidence need to be available.
3. Articles reporting primary research and traditional literature reviews To some extent?
are accessible to practitioners. For many reasons systematic reviews
may not be available or produced in an area of interest. When this is
the case, primary research and traditional reviews published in
journals can be used.
4. ‘‘Cutting-edge’’ practices, panaceas, and fashionable new ideas are To some extent?
treated with healthy skepticism. One characteristic of areas of
practice that are not particularly evidence based is the constant
search for and promotion of the newest solution or cure. This
characteristic is found in popular books on topics such as dieting,
self-help, and indeed management.
5. There is a demand for evidence-based practice from clients and To some extent?
customers. If the clients or customers of a particular practice do not
want or even reject evidence-based practice then it is simply
impossible to practice in an evidence-based way.
6. Practice decisions are integrative and draw on the four sources of To some extent?
information and evidence described above. As emphasized, (Difficult to judge.)
evidence-based decision making is more than looking at external
published evidence. Rather, it is about combining evaluated external
evidence, the perspectives of those who may be affected by the
decision, information from the local context, and practitioner
experience and expertise.
7. Initial training and continuing professional development (CPD) adopt To a very limited
evidence-based approaches. From an evidence-based perspective, extent?
initial training and CPD focus on developing evidence-based
approaches to practice. This involves training practitioners to identify
and critically evaluate external and contextual evidence relevant to a
specific practice problem to help inform a practice decision. This
approach creates an active need to obtain and use relevant evidence,
as it is being used directly to help solve a problem.

annual conference did have EBMgt as systematic reviews in I–O psychol-


one of its themes. ogy. SIOP is currently developing its
2. Systematic reviews are produced and Science You Can Use series, which
made accessible. Although we have will contain reviews that are not
plenty of traditional reviews and systematic as such but will go some
meta-analyses, there are very few of the way to summarizing research
10 R.B. Briner and D.M. Rousseau

findings that I–O practitioners can else will. However, as I–O psycholo-
use. gists, we may at the same time try
3. Articles reporting primary research to rework the fad into something
and traditional literature reviews are closer to our own practice and to
accessible to practitioners. We have established and evidence-based tech-
found that this is a difficulty for niques.2
many I–O psychology practition- 5. There is a demand for evidence-based
ers unless they are still attached in practice from clients and customers.
some way to a university. Although Many of our major clients are those
abstracts are easily accessible, pur- working at mid to senior levels in
chasing single articles from publishers HRM. HRM is not a field that has
can be costly. In addition, to build embraced the notion of evidence-
up even limited knowledge of the based practice in any significant way.
evidence in a particular area can Although, of course, managers do
require access to dozens of arti- not actively seek to purchase inef-
cles, which may be prohibitively fective I–O psychology products or
expensive. It also appears to be services, they are under pressure to
the case that not all I–O psychol- meet certain shorter term goals. They
ogy practitioners, depending some- may therefore come to depend on
what on where they trained, are a general impression that particu-
highly skilled in reading and digesting lar products or techniques ‘‘work’’
(sometimes rather indigestible) jour- rather than whether they will work
nal articles. in their specific context given the
4. ‘‘Cutting-edge’’ practices, panaceas, problem they are trying to solve.
and fashionable ‘‘new’’ ideas are HRM departments may also lean on
benchmarking or mimicry by adopt-
treated with healthy skepticism. As
ing the same I–O psychology prac-
a consequence of our training as psy-
tices already used by their more
chologists, we should be inclined
successful competitors. The authors
to be quite skeptical or at least
have also heard many times from I–O
are inclined to ask about evidence
psychologists who say they wish to
and empirical support. At the same
practice in an evidence-based way
time, however, we are also somewhat
but that clients have often already
drawn to what might be called man-
decided what they want (e.g., assess-
agement fads and fashions. Some of
ment centers, training programs, and
the recent focal articles in this jour-
employee attitude surveys) and are
nal demonstrate that we are some- asking the I–O psychologist as a
times inclined to pick up and run technical specialist to deliver it. This
with the Next Big Thing even where situation suggests that our clients are
evidence does not yet exist or is ques- not demanding an evidence-based
tionable. In addition, next time you approach in the sense discussed here,
attend the SIOP annual conference, although they are of course interested
check out the products and services in adopting practices they believe to
for sale in the exhibit hall. In our be effective.
experience, many of these feel more 6. Practice decisions are integrative and
like fads than evidence-based inter- draw on the four sources of informa-
ventions (and often no supporting tion and evidence described above.
evidence is presented). One reason
we pay attention to fads is that clients
often demand the latest thing, and if 2. We thank an anonymous reviewer for these
we don’t deliver it, then some one observations.
Evidence-based I–O psychology 11

This is the most difficult character- passive way to learn and retain infor-
istic of evidence-based practice to mation. We note that in both the
assess without access to numerous United States and Britain, the major-
observations of what practicing I–O ity of practicing I–O psychologists
psychologists actually do. In addition, have terminal master’s degrees. The
it may be that as discussed above, we typical master’s program in the field
are not the decision makers; instead, has no required supervised intern-
we play an advisory role, provid- ships and does not train its stu-
ing information and interpretations to dents to conduct literature searches
the decision makers. However, tak- on practice topics let alone system-
ing each source of information in atic reviews. The forms of CPD used
turn, I–O psychologists do, as dis- by SIOP and other I–O psychol-
cussed above, draw to some extent ogy professional bodies also tend to
on evaluated external evidence when be fairly traditional. In fact, some
making decisions, even though few of these forms of CPD reverse the
systematic reviews are available and approach adopted by evidence-based
access to primary research can be dif- practice in that they present par-
ficult. The perspectives of those who ticipants with recent research find-
may be affected by the decision are ings, evidence, or new techniques
likely to be taken into account at and then discuss how they might
least to some extent because of the be used in practice rather than start-
APA’s Ethics Code (2002; assuming
ing with practice problems and then
we are complying with the code)3 and
searching for and evaluating the
also because of the broader aware-
evidence that may help solve the
ness we should have as psychologists
problem.
about our responsibilities to organi-
zational and individual clients. We
So, what is it possible to conclude from
are likely to look for and use evi-
this analysis of the extent to which I–O
dence from the local context and
psychology shares some of the character-
attempt some initial assessment of
istics of evidence-based practice? First, we
the problem or situation and seek
out organizational data that might suggest that I–O psychology is not strongly
help with problem diagnosis. The evidence based in the sense that the term
use of practitioner experience and is used in other professions. But, we can
judgment seems highly likely, partic- say with some confidence that we are as
ularly if the problem or technique is a profession extremely well positioned to
one we have encountered frequently adopt, should we wish to do so, many
before. of these characteristics. Next, I–O psy-
7. Initial training and continuing pro- chologists in many instances are not the
fessional development (CPD) in evi- key decision makers but, rather, sources of
dence-based approaches. Training in information and advice to managers mak-
I–O psychology master’s degrees and ing the decision (see below). Last, it is
doctorates tends to be of the fairly clear that there are many barriers to the
traditional academic variety where adoption of evidence-based practice, some
students are expected in a relatively within and others outside our control. Hav-
ing evaluated I–O psychology as falling
somewhat short on evidence-based prac-
3. Ethical I–O psychology and evidence-based I–O tice and supports for it, we turn to two
psychology are similar in other respects including important means for bridging I–O psychol-
the focus on being concerned and explicit about
the evidence for the benefits of interventions and ogy’s own research–practice gap: practice-
the evidence that interventions are not harmful. oriented evidence and systematic reviews.
12 R.B. Briner and D.M. Rousseau

Key Strategies for Promoting Anderson et al. (2001) noted that between
Evidence-Based Practice 1949 and 1965, practitioners authored a
full 36% of JAP articles (31% by prac-
I–O psychologists are not one seamless
titioners alone). From 1990 to 2000 (the
community of comembers of the same
terminal year of their survey), practitioners
discipline. Practitioners and scholars in I–O
authored only 4% of JAP articles (1% by
psychology are largely distinct communities
practitioners alone). The other I–O journals
of interest, knowledge, and social ties.
manifest a similar decline. Many factors
In promoting evidence-based practice, it
may account for this decline in practitioner
is advantageous in such circumstances
research publication in our field’s journals,
to design ways of communicating and
including a shift in journal focus to more
sharing ideas that serve the interests of
academic topics of rigor, greater corporate
both. Although evidence-based practice
concern for protecting intellectual property,
involves the better use and integration of
as well as ramped-up global competition
evidence and information from all four
and its accompanying time and resource
sources described above, we focus here
crunch, which in turn limited practitioner
on improving the use of critically evaluated
opportunity for research let alone publi-
research evidence.
cation. One conclusion is apparent: I–O
Between the research and practice psychology’s academics and practitioners
domains of I–O psychology, we need are not mingling with each other in our jour-
devices for translating back and forth infor- nals. Regardless of the underlying reasons,
mation and knowledge, promoting better there is one serious consequence of the
quality communication and learning. We decline in practitioner conducted research
note that in the SIOP practitioner survey publication: Academics are the ones asking
described above, a frequent practitioner the research questions and interpreting the
request to SIOP was for clarification of answers.
standards for I–O practice and better ways If the gap between research and prac-
of differentiating I–O psychologists from tice in I–O psychology is at least partly
other practitioners in the marketplace. Such attributable to lower participation by prac-
clarification and professional differentiation titioners in research, the problem may
can come from creating the evidence- be exacerbated by omission in current
oriented products and associated processes research of the kinds of complex prob-
proposed here. Such products can simulta- lems in complex settings faced by prac-
neously meet the needs of I–O psychology’s titioners in their work. An antidote to
practitioners and scholars, adding value the latter has been suggested in the form
to both. The two products we suggest as of engaged scholarship (Van de Ven,
critical to evidence-based practice in I–O 2007) and scholar–practitioner collabora-
psychology are practice-oriented evidence tion (Lawler & Mohrman, in press), where
and systematic reviews. academics and practitioners work together
to formulate research questions, investigate
them, and draw conclusions.
Practice-Oriented Evidence
If indeed I–O psychology research is
Most research published in I–O psychol- now academic centric, the gap between
ogy’s premier journals, including Journal research and practice entails problems in
of Applied Psychology (JAP ), Person- knowledge transfer. It takes two to tango,
nel Psychology, and Journal of Occupa- and difficulties transferring knowledge can
tional and Organizational Psychology is be because of communication issues on
theory-oriented investigations authored by both sides (what’s readable, accessible,
academy-based I–O psychologists answer- understandable, and interesting?). Barriers
ing questions of interests to other aca- to transfer may also reside in the nature of
demics. This was not always the case. the knowledge itself. As a case in point,
Evidence-based I–O psychology 13

meta-analysis and literature reviews in I–O average and college reputation. Another
psychology have led to the formulation reason for caution in using tests of GMA is
of general knowledge principles based on concern over adverse impact, even though
scientific evidence (The Handbook of Prin- some methods have relatively low adverse
ciples of Organizational Behavior: Indis- impact (Schmidt, 2009).
pensable Knowledge for Evidence-Based This debacle may be tied to what Van
Management ; Locke, 2009). The Handbook de Ven and Johnson (2006) refer to as
has over 30 chapters summarizing several a knowledge production problem, argu-
hundred evidence-based principles, the vast ing that the research–practice gap is best
majority of them derived from I–O psy- bridged by producing practice-oriented sci-
chology research and all intended to guide entific knowledge via research approaches
the practice of current and future managers engaging both academics and researchers
and other practitioners. This book provides collaboratively. This approach calls for
many examples and is written in plain combining the knowledge of practition-
English. It represents, perhaps at its best, ers and the knowledge of academics at
knowledge (principles) based on what might all stages of the research process. This
be called ‘‘researcher-oriented evidence,’’ knowledge production problem has been
that is, evidence from rigorous tests of the- encountered, and addressed, in other
ory, replicated and found to be relatively evidence-based practice fields.
generalizable over time and context. Medical researchers and clinicians dis-
So, what’s wrong with that? In many tinguish between two kinds of research
ways, these research-based principles evidence: disease oriented and patient ori-
achieve our ideal as scientists: to under- ented. ‘‘Disease-oriented evidence’’ (DOE)
stand the world well and disseminate this focuses on the causes of disease providing
knowledge. The downside is this: It is evidence of pathology and ways of altering
not always obvious to practitioners, cer- the condition (e.g., drugs and surgery). In
tainly not to the least experienced or less I–O psychology, our version of DOEs might
reflective, how exactly to apply the prin- take the form of phenomena-oriented evi-
ciples identified in such research. Let’s dence, such as the origins of job stress or job
take the classic example of the finding satisfaction. The second kind of research
that general mental ability (GMA) is pos- evidence in medicine is ‘‘patient-oriented
itively related to individual performance evidence that matters’’ (POEMs), evidence
(Schmidt, 2009). A well-established finding gathered from studies of real patients about
over several decades, one practice impli- issues such as mortality, morbidity, and
cation is that if an organization seeks to quality of life. An example of a POEM
improve the quality of its workforce and is a study comparing one antihyperten-
the performance of individual members, it sive drug to another to determine which
should select on intelligence. For a host of reduced mortality from hypertension over a
reasons, this principle is not widely pro- 10- to 20-year period. In I–O psychology,
moted by I–O psychology practitioners and our version of POEMs might take the form
is soundly rejected by even experienced HR of studies contrasting two interventions to
managers (Rynes et al., 2002). Practitioners reduce job stress that assess the types of
think about the educated idiot who is book individuals, work settings, and job strains
smart, tests fantastically well, and can’t they best ameliorate.
match socks. Managers fear being labeled There is a growing trend in the prac-
elitist and perhaps wonder whether they tice of medicine to value patient-oriented
would have gotten their own jobs if their data more highly than DOE. However,
company used IQ tests. Or, they use a test because practice-oriented evidence does
of GMA and find that it doesn’t improve per- not yet exist to inform every clinical
formance levels over the biographical infor- need, practitioners must use other ways
mation they already rely on like grade point of making decisions too, including relying
14 R.B. Briner and D.M. Rousseau

on their knowledge of basic physiological of GMA. Practice-oriented research could


processes. We expect much the same look into whether performance criteria in
alternative forms of decision making in use affected the value and usefulness practi-
an evidence-informed practice of I–O tioners attach to indicators of GMA. Perhaps
psychology. settings where innovation and creativity are
An example of practice-oriented evi- important performance metrics place more
dence in I–O psychology are the reports by value on mental ability than those where
Robert Pritchard and his team, developing more routine performance is involved. Aca-
and investigating the use of the Produc- demically oriented evidence indicates that
tivity Measurement and Enhancement Sys- GMA is likely to predict performance in
tem (ProMES) system for job analysis and either case. Practitioners may only find
strategic planning (e.g., Pritchard, Harrell, GMA useful where mental ability is an
DiazGranados, & Guzman, 2008). Differ- organizationally valued contributor to per-
ences identified between studies in how formance.
thoroughly the ProMES system was applied
suggested that several implementation-
related factors, including the extent users
adhered to the ProMES process and the
Systematic Reviews
quality of the feedback provided, affected Systematic reviews are fundamental to
the overall productivity gains associated evidence-based practice. As such, much is
with ProMES. Pritchard and colleagues then written about them in other fields (Gough
address the circumstances under which et al., 2003; Soukup, 2000; Tranfield,
there are differences in implementation or Denyer, & Smart, 2003). Experience indi-
compliance with standard practices and cates that it is impossible to fully engage
the sensitivity of outcomes to these vari- in evidence-based practice without them:
ations (see Pritchard et al., 2008). In the Such reviews provide one of the four
context of widespread variation in orga- sources of information required when mak-
nizational and management practice (from ing evidence-based decisions. Applied to
performance appraisals to quality programs) the example in Table 1, a systematic
as well as in individual implementers, evi- review’s purpose would be to search for,
dence regarding the effects of such variabil- collect, critically appraise, and pull together
ity on outcomes has considerable practical research evidence relevant to the causes
and scholarly value. and possible solutions to the problem
Consider what practice-oriented evi- of high absence. I–O psychologists are
dence might mean for some of the stickier
avid producers and consumers of literature
problems in I–O psychology. We know
reviews. However, systematic reviews are
that GMA is predictive of individual per-
very different from nearly all those pub-
formance, but organizations are reluctant
lished in I–O psychology.
to accept or act on this knowledge (Rynes
et al., 2002), often preferring intuitive selec-
tion methods (Highhouse, 2008). Practice- Systematic reviews are literature reviews
oriented evidence could be developed from that adhere closely to a set of scientific
investigations into conditions making use methods that explicitly aim to limit sys-
of GMA as a selection criterion more read- tematic error (bias), mainly by attempting
ily useful. Looking into conditions of use to identify, appraise and synthesize all
could identify, for example, whether the relevant studies (of whatever design) in
backlash is to written IQ-type tests, where order to answer a particular question (or
structured interview questions that tapped set of questions). In carrying out this task
GMA would be more acceptable, or how they set out their methods in advance,
the concerns over adverse impact could be and in detail, as one would for any piece
better balanced with the predictive validity of social research. In this respect . . . they
Evidence-based I–O psychology 15

are quite unlike most ‘‘traditional’’ narra- would attempt to find evidence about the
tive reviews (Petticrew & Roberts, 2006, relative effectiveness of different forms of
pp. 9–10). absence management interventions given
the current and desired absence rates and
Systematic reviews are, essentially, a taking into account as much as possible
way of analyzing existing research using aspects of the context such as the type of
explicit and replicable methods, allowing employees involved, the sector, and the
conclusions to be drawn about what is existing work arrangements and absence
known and what is not known in relation policies. In the context of evidence-based
to the review question (and within the practice, systematic reviews can take forms
limitations of the method). Similar, but not akin to phenomena-oriented evidence or
identical to the traditional meta-analysis, practice-oriented evidence, depending on
systematic reviews are studies of studies. the review questions and their intended
Meta-analyses are a type of systematic use as well as the kinds of research avail-
review but one that uses only quantitative able. In evidence-based fields, an important
data and statistical synthesis and focuses result of systematic reviews is guidelines for
on a question repeatedly addressed in the practice.
same way by researchers rather than a Systematic reviews can be useful for
practice question or problem. As with meta- purely academic research purposes too.
analyses, systematic reviews are conducted We may, for example, be interested in
out of recognition that single empirical collecting all available evidence about
studies, although useful and sometimes absence-management interventions to pro-
informative, should not be emphasized
vide a more general overview about what
because their biases and limitations cannot
is known, and not known, about the effi-
be fully accounted for. Looking at all
cacy of such interventions. In this respect,
relevant studies, systematically gathered,
a systematic review might differ from the
constitutes more reliable evidence.
traditional meta-analysis in that it would
Thus, in the context of evidence-
also consider qualitative information and
based practice, neither traditional literature
descriptions, not being limited to effect
reviews nor meta-analyses are especially
sizes alone depending on the review ques-
useful. First, traditional literature reviews
are open to many forms of bias. For tion (Rousseau, Manning, & Denyer, 2008).
example, reviewers do not make clear how All systematic reviews follow a process of
they have selected the studies they have clearly specified stages. One example of this
included, do not critically appraise them process (adapted from Petticrew & Roberts,
in an explicit or systematic way, and do 2006) contains seven stages.
not usually pull them together or synthe-
size findings across studies. Second, tra- 1. Identify and clearly define the ques-
ditional reviews do not usually focus on tion the review will address. The ques-
a specific research, practice question, or tion needs to be sufficiently specific so
problem. It is this latter point that also that it is clear, in principle, what types
differentiates a systematic review from the of data would be relevant. Aspects of
quantitative meta-analysis used tradition- the context (e.g., population, sector,
ally in I–O psychology. The process of and organizational type), the inter-
making evidence-based decisions requires ventions (what qualifies as a relevant
more focused and tailored reviews of evi- intervention?), the mechanisms link-
dence where both a practice question or ing intervention to outcomes (e.g.,
problem and the conditions to which the processes, mediators, and modera-
evidence might be applied are taken into tors), and the outcomes themselves
account. Returning to the case of high (which data are the outcomes of inter-
absence in Table 1, a systematic review est) are also clearly specified.
16 R.B. Briner and D.M. Rousseau

2. Determine the types of studies and Synthesis may be quantitative, qual-


data that will answer the question. itative, or both. Review findings are
The criteria used to decide which often described in terms of the overall
studies will be selected or excluded number of studies found, the quality
from the review are identified. For profile of this group of studies, and
example, a review that addresses a the number of studies that obtained
causal question might exclude studies particular results.
with cross-sectional designs. The aim 7. Disseminate the review findings. A
is to increase the chances that all full report of a systematic review can
relevant studies are included and all be quite large. In addition, shorter
those that are irrelevant are excluded. journal article length versions or even
3. Search the literature to locate relevant shorter summaries may be produced.
studies. A clear search strategy is used Dissemination is often planned at the
specifying, for example, key words, outset of a systematic review given
the databases to be searched, and the aim is often to inform practice.
how, and whether unpublished data
will be found and included. Although systematic review is rare in
4. Sift through all the retrieved studies to I–O psychology at present, I–O psychol-
identify those that meet the inclusion ogists are certainly familiar with its general
criteria (and need to be examined approach. The underlying logic of system-
further) and those that do not and atic review is similar to that of many psy-
chological research methods and is similar
should be excluded. Each study is
to meta-analyses. Meta-analysis’ exclusive
examined, usually by two review
use of quantitative data and statistical rather
team members, and checked against
than other forms of synthesis sets it apart
the inclusion and exclusion criteria.
from a systematic review, which often uses
Where agreement is not possible, a
different types of data and different forms
third reviewer assesses the study. At
of synthesis. In addition, meta-analyses
this stage, it is not uncommon to find
can only address questions that have been
that only a fraction of the initial pool
addressed many times in more or less the
of studies can be included. same way by researchers (e.g., the corre-
5. Critically appraise the studies by lation between job satisfaction and per-
assessing the study quality deter- formance) rather than questions that arise
mined in relation to the review ques- from practice problems, where an array of
tion. The quality of each study is data types may be required to formulate an
critically appraised or evaluated in answer.
relation to the review question as An example of a structured abstract from
research quality can only be judged in a systematic review particularly relevant to
relation to the question. Even where I–O psychology is presented in Table 3.
studies meet the inclusion criteria, This demonstrates much of the underly-
they are still likely to vary in terms ing methodology and shows the explicit
of quality. Assessing quality allows and systematic nature of the method.
the review conclusions to clearly state Joyce, Pabayo, Crichley, & Bambra’s (2010)
how many of the studies included review clearly states the review objectives,
were, for example, of very high, search strategy, criteria for including stud-
medium, and low quality. ies, method of analysis (in this case, a
6. Synthesize the findings from the stud- narrative synthesis as the studies were dis-
ies. A key part of any systematic similar), the number of studies found, and
review is the pulling together of find- the findings of each.
ings from across the studies to repre- A few aspects of systematic reviews
sent what is known and not known. and their differences from traditional I–O
Evidence-based I–O psychology 17

Table 3. Example of a Systematic Review Abstract

Flexible working conditions and their effects on employee health and well-being (Joyce et al.,
2010)
Background: Flexible working conditions are increasingly popular in developed countries,
but the effects on employee health and well-being are largely unknown
Objectives: To evaluate the effects (benefits and harms) of flexible working interventions on
the physical, mental, and general health and well-being of employees and their families
Search strategy: Our searches (July 2009) covered 12 databases including the Cochrane
Public Health Group Specialized Register, CENTRAL, MEDLINE, EMBASE, CINAHL,
PsycINFO, Social Science Citation Index, ASSIA, IBSS, Sociological Abstracts, and
ABI/Inform. We also searched relevant Web sites, hand searched key journals, searched
bibliographies, and contacted study authors and key experts
Selection criteria: Randomized controlled trials, interrupted time series, and controlled
before and after studies (CBA), which examined the effects of flexible working
interventions on employee health and well-being. We excluded studies assessing
outcomes for less than 6 months and extracted outcomes relating to physical, mental, and
general health/ill-health measured using a validated instrument. We also extracted
secondary outcomes (including sickness absence, health service usage, behavioral
changes, accidents, work–life balance, quality of life, health and well-being of children,
family members, and coworkers) if reported alongside at least one primary outcome
Data collection and analysis: Two experienced review authors conducted data extraction and
quality appraisal. We undertook a narrative synthesis as there was substantial
heterogeneity between studies
Main results: Ten studies fulfilled the inclusion criteria. Six CBA studies reported on
interventions relating to temporal flexibility: self-scheduling of shift work (n = 4), flexitime
(n = 1), and overtime (n = 1). The remaining four CBA studies evaluated a form of
contractual flexibility: partial/gradual retirement (n = 2), involuntary part-time work
(n = 1), and fixed-term contract (n = 1). The studies retrieved had a number of
methodological limitations, including short follow-up periods, risk of selection bias, and
reliance on largely self-reported outcome data. Four CBA studies on self-scheduling of
shifts and one CBA study on gradual/partial retirement reported statistically significant
improvements in either primary outcomes (including systolic blood pressure and heart rate;
tiredness; mental health, sleep duration, sleep quality, and alertness; and self-rated health
status) or secondary health outcomes (coworker social support and sense of community),
and no ill-health effects were reported. Flexitime was shown not to have significant effects
on self-reported physiological and psychological health outcomes. Similarly, when
comparing individuals working overtime with those who did not, the odds of ill-health
effects were not significantly higher in the intervention group at follow-up. The effects of
contractual flexibility on self-reported health (with the exception of gradual/partial
retirement, which when controlled by employees improved health outcomes) were either
equivocal or negative. No studies differentiated results by socioeconomic status, although
one study did compare findings by gender but found no differential effect on self-reported
health outcomes
Authors’ conclusions: The findings of this review tentatively suggest that flexible working
interventions that increase worker control and choice (such as self scheduling or
gradual/partial retirement) are likely to have a positive effect on health outcomes. In
contrast, interventions that were motivated or dictated by organizational interests, such as
fixed-term contract and involuntary part-time employment, found equivocal or negative
health effects. Given the partial and methodologically limited evidence base, these
findings should be interpreted with caution. Moreover, well-designed intervention studies
are needed to delineate the impact of flexible working conditions on health, well-being,
and health inequalities
18 R.B. Briner and D.M. Rousseau

psychology reviews are worth noting. First, together is clear to the reader, as the findings
the exact methods used by the reviewers to of each study are extracted and presented.
find, select, and exclude studies are open to Fourth, and again unlike traditional
examination and scrutiny. Readers are thus reviews, systematic reviews also allow us to
able to reach their own judgments about the identify the quantity and quality of studies
efficacy and appropriateness of the method and many aspects of study heterogeneity
and also how much confidence they can (method, population, design, findings, etc.).
place in the findings. This information is essential if we want
Second, whether we are I–O practition- to draw conclusions about what is known
ers, researchers, or both, as we have little and not known in relation to the review
access to systematic reviews, we tend to question, the basis of these claims, and the
develop an implicit though somewhat inac- confidence with which they can be made.
curate sense of the quantity and quality In addition to full-blown systematic
of research about a particular technique reviews of the sort described in Table 2,
or intervention. We tend to assume that there are other quicker and more focused
the research is ‘‘out there’’ somewhere and ways of doing systematic reviews that share
have vague recollections of particular stud- many of their qualities. Although they may
ies and their findings. As mentioned, one not be as thorough and therefore not as
common finding from systematic reviews informative as systematic reviews, they
is that they reveal far fewer studies directly can still provide important information and
relevant to a given question than commonly practice insights, especially where time is
assumed. In the case of the flexible working of the essence. These go by various names,
review by Joyce et al. (2008), the large num- including rapid evidence assessment and
ber of possibly relevant studies that were best evidence topics and can be completed
identified was substantially reduced as the more quickly by restricting the parameters
review team applied their inclusion criteria: in various ways, such as using fewer search
terms, using a smaller date range, and
• 11,954 articles initially identified as searching in fewer databases or across fewer
possibly relevant from key word journals.
searches on databases, citation sear- We believe that systematic reviews can
ches, hand searching, and contacting be and will become an essential part of
experts; the I–O psychologist’s toolkit. Not only do
• 11,740 articles then excluded on basis they allow practitioners to provide more
of title and/or abstract; evidence-based advice and share with their
• 214 full-text articles obtained and then clients and customers the basis of that
screened using inclusion criteria; and advice, they also allow researchers in a
• 10 studies included in the final review. more structured way to identify important
gaps in knowledge. Also importantly,
As we will discuss further below, one systematic reviews can highlight where
hallmark of being an evidence-based prac- conducting more research on the same
titioner or researcher is having a quite well- question using a particular method is
developed, specific, and explicit awareness unlikely to yield any new information.
of what evidence is ‘‘out there,’’ the quality
of that evidence, and what, when taken as
Barriers to Evidence-Based
a whole, it might mean.
Practice in I–O Psychology
Third, unlike traditional reviews, the pro-
cesses through which reviewers reach their There are numerous barriers to the adoption
conclusions are explicit and transparent. of evidence-based practice and numerous
Whether the synthesis is statistical or narra- ways in which they can be overcome.
tive (as in Joyce et al., 2008), the basis on Here, we focus on just a few. First is
which the data were summarized or pulled the apparent lack of demand from our
Evidence-based I–O psychology 19

clients for evidence-based I–O psychology. see it as a sign of competence. I–O


It is readily apparent from looking at the psychologists are especially challenged
rapid adoption of some new and ‘‘cutting- because other organizational consultants
edge’’ practices that such decisions are outside the discipline can be extremely
not made in an evidence-based way. If bullish about their products and services
individual managers are mostly rewarded despite their own absence of evidence.
for achieving short-term goals as fast as As skepticism is a key ingredient of
possible rather than doing what works in the evidence-based practice, its limited pop-
longer term, why would they be interested ularity in I–O psychology is something of
in evidence-based practice? Perhaps, the a barrier. One way this can be overcome
only way to overcome this barrier is by is to remind ourselves that skepticism is
working with organizations to demonstrate fundamental to scientific inquiry and to
that approaching problems in an evidence- any area of practice based on science. It
based way is more likely to produce also has the potential to clearly differentiate
effective and sustainable solutions. It is us from other organizational practitioners
also important to emphasize that evidence- and consultants, particularly if the disci-
based practice constitutes a family of pline supports practice through systematic
approaches to making decisions and is not reviews, evidence-based guidelines, and
intended to provide the answer to every practice-oriented research that contribute
problem but rather improve the process and to more effective practice.
outcome of decision making. A fourth barrier concerns the politics of
A second barrier is the predominance evidence in organizations. Power and poli-
of master’s-level practitioners who have tics are fundamental to decision making and
learned to practice I–O psychology in unsu- also surround the identification and use of
pervised ways. Because of the level of such evidence in organizations. Senior leaders
programs, such practitioners have a lim- may feel they have the right or even respon-
ited understanding of research, a limited sibility to make decisions based on their
capacity to access new evidence, and lack experience and judgment that seem to fly
the skills to conduct their own systematic in the face of the available evidence. The
reviews let alone primary research. Collec- need to be explicit in evidence-based deci-
tively, we currently do not have enough sion making means that those with vested
of the necessary skills to widely under- interests in a particular course of action may
take evidence-based practice, even though find it more difficult to hide such interests.
our background as psychologists gives us In general, an evidence-based approach
a strong foundation on which to build. may prove challenging particularly in orga-
CPD that enables evidence-based practice nizations with highly political cultures.
and helping I–O practitioners to access Although it is impossible to remove pol-
research evidence are two possible ways itics from evidence and decision making,
to overcome this barrier. It may also help to evidence-based approaches do at least offer
increase the opportunities for those qual- the possibility of making clearer distinctions
ified at the master’s level to go on to among politics, values, interests, and other
complete doctorates. forms of information such as research evi-
As a consequence of our desire to market dence. The more decision makers are held
I–O psychology, we may be reluctant accountable for their decisions, the more
to acknowledge the limitations of our likely they are to welcome such distinctions.
knowledge where evidence is mixed or
where there are grounds for uncertainty.
Prospects for Evidence-Based
It can be difficult to achieve a balance
I–O Psychology
between promoting a profession while at
the same time acknowledging its limitations This article concludes that I–O psychol-
as clients may find this unnerving or ogy cannot yet claim to be fully evidence
20 R.B. Briner and D.M. Rousseau

based. Our analysis suggests that we have simplest and least informative analyses such
some way to go before we can reason- as cross tabs and zero-order correlations.
ably claim to be an evidence-based area I–O psychologists should be able to work
of research and practice in the sense it with organizations to help them get the
is used in other areas of evidence-based most out of the data they already have and
practice. At the same time, I–O psychol- where appropriate suggest additional data
ogists are uniquely qualified to undertake collection to develop a fuller picture of what
the key activities required by evidence- is happening in the organization.
based practice such as evaluating exter- Our roles as organizational knowledge
nal research evidence and collecting and brokers as well as our own evidence-based
using internal evidence for organizational practice would be facilitated by the use
diagnosis. In evidence-based practice, I–O of systematic reviews on practice ques-
psychologists as trainers and consultants, tions. Systematic reviews need to be part
in-house management and staff, are posi- of the professional training of I–O psychol-
tioned to enable an array of approaches. ogists. A professional activity accessible to
These include, various combinations as both master’s-prepared and doctoral-level
warranted, using evidence-based processes I–O psychologists, systematic reviews, and
for making decisions, giving feedback, as briefer versions are important ways of help-
well as incorporating evidence-based con- ing would-be practitioners learn to gather
tent, that is, research findings, into their and use evidence while at the same time
decisions and practices (Briner et al., 2009). developing their skills in formulating ques-
Beyond making our own practice more tions, structuring reviews, and synthesizing
evidence based, we also have an important findings.
part to play in helping organizations Third, I–O psychologists have the back-
and those who manage them become ground to enable us to work as facilitators
more evidence based. We envisage this and coaches for managers and management
could happen in several ways. First, we teams seeking to engage in evidence-based
could provide systematic review services management as well as helping organi-
(or their briefer and quicker versions) to zations collect the external and internal
managers and organizations. These might evidence they may need (described above).
be seen as particularly useful where the We can also help collect information about
organization is trying to decide whether the perspectives of those who may be
to invest a large sum of money in a affected by a decision and help make
particular intervention or program. For a explicit managers’ own expertise and expe-
fraction of the proposed program budget, rience and how it is shaping a decision.
we would be able to provide an objective In effect, we can support organizations to
and reasonably comprehensive review of make decisions in a conscientious, explicit,
what the published and (where available) and judicious way—in short, to help orga-
unpublished research tells us about how nizations to practice EBMgt.
effective the intervention is known to be Evidence-based practice is not about
and whether it might work here. perfection. Rather, it is about on-going
Second, we have the skills to help pursuit of mindful improvement in our
organizations either make sense of their uptake and use of both scholarly research
existing internal data or collect new data and practice-related facts and assessments.
that might diagnose problems or show We hope this article and the responses that
what is working and why. We believe it follow help move I–O psychology in an
is not unusual, even where organizations evidence-based direction.
have large quantities of data about, for
example, employee behavior, performance, References
and attitudes, for those data to remain Anderson, N., Herriot, P., & Hodgkinson, G. P. (2001).
largely unanalyzed or subject only to the The practitioner-researcher divide in industrial,
Evidence-based I–O psychology 21

work and organizational (IWO) psychology: Where Learmonth, M., & Harding, N. (2006). Evidence-based
are we now and where do we go from here? Journal management: The very idea. Public Administration,
of Occupational and Organizational Psychology, 84, 245–266.
74, 391–411. Locke, E. A. (2005). Why emotional intelligence is
Briner, R. B. (1998). What is an evidence-based an invalid concept. Journal of Organizational
approach to practice and why do we need one Behavior, 26, 425–431.
in occupational psychology? Proceedings of the Locke, E. A. (2009). Handbook of principles of
1998 British Psychological Society Occupational organizational behavior: Indispensable knowledge
Psychology Conference, pp. 39–44. for evidence-based management (2nd ed.). New
Briner, R. B., Denyer, D., & Rousseau, D. M. (2009). York: Wiley.
Evidence-based management: Concept clean-up Macey, W. H., & Schneider, B. (2008). The meaning
time? Academy of Management Perspectives, 23, of employee engagement. Industrial and Organi-
19–32. zational Psychology: Perspectives on Science and
Cherniss, C. (2010). Emotional intelligence: Toward Practice, 1, 3–30.
clarification of a concept. Industrial and Organi- Mayer, J. D., & Salovey, P. (1997). What is emo-
zational Psychology: Perspectives on Science and tional intelligence? In P. Salovey and D. Sluyter
Practice, 3, 110–126. (Eds.), Emotional development and emotional intel-
Friedson, E. (1986). Professional powers: A study of the ligence: Educational implications (pp. 3–34). New
institutionalization of formal knowledge. Chicago: York: Basic Books.
University of Chicago Press. McCall, M. W. (2010). Recasting leadership develop-
Goleman, D. (1998). Working with emotional intelli- ment. Industrial and Organizational Psychology:
gence. New York: Bantam Books. Perspectives on Science and Practice, 3, 3–19.
Gough, D. A., Kiwan, D., Sutcliffe, S., Simpson, D., McDaniel, M. A., Whetzel, D. L., Schmidt, F. L., &
& Houghton, N. (2003). A systematic map and Maurer, S. D. (1994). The validity of employment
synthesis review of the effectiveness of personal interviews: A comprehensive review and meta-
development planning for improved student learn- analysis. Journal of Applied Psychology, 79,
ing. Retrieved from http://eppi.ioe.ac.uk/EPPIWeb 599–616.
Content/reel/review_groups/EPPI/LTSN/LTSN_ McKenna, D. D., & Davis, S. L. (2009). Hidden in
June03.pdf plain sight: The active ingredients of executive
coaching. Industrial and Organizational Psychol-
Hambrick, D. C. (1994). What if the academy actually
ogy: Perspectives on Science and Practice, 2,
mattered? Academy of Management Review, 19,
244–260.
11–16.
Merrill, D. C., Reiser, B. J., Merrill, S. K., & Landes, S.
Highhouse, S. (2008). Stubborn reliance on intuition
(1995). Tutoring: Guided learning by doing.
and subjectivity in employment selection. Indus-
Cognition and Instruction, 13, 315–372.
trial and Organizational Psychology: Perspectives
Morrell, K. (2008). The narrative of ‘‘evidence based’’
on Science and Practice, 1, 333–342.
management: A polemic. Journal of Management
Hollenbeck, G. (2009). Executive selection—What’s
Studies, 45, 613–635.
right . . . and what’s wrong. Industrial and Organi- Murphy, K. R. (2008). Explaining the weak relationship
zational Psychology: Perspectives on Science and between job performance and ratings of job perfor-
Practice, 2, 130–143. mance. Industrial and Organizational Psychology:
Joyce, K., Pabayo, R., Critchley, J. A., & Bambra, C. Perspectives on Science and Practice, 1, 148–160.
(2010). Flexible working conditions and their Petticrew, M., & Roberts, H. (2006). Systematic
effects on employee health and wellbeing. reviews in the social sciences: A practical guide.
Cochrane Database of Systematic Reviews, Art. No. Oxford: Blackwell.
CD008009. doi: 10.1002/14651858.CD008009. Pfeffer, J., & Sutton, R. I. (2006). Evidence-based man-
pub2. agement. Harvard Business Review, 84, 62–74.
Judge, T. A., Heller, D., & Mount, M. K. (2002). Five Pritchard, R. D., Harrell, M. M., DiazGranados, D., &
factor model of personality and job satisfaction: A Guzman, M. J. (2008). The productivity measure-
meta-analysis. Journal of Applied Psychology, 87, ment and enhancement system: A meta-analysis.
530–541. Journal of Applied Psychology, 93, 540–567.
Judge, T. A., & Illies, R. (2002). Relationship of Reay, T., Berta, W., & Kohn, M. (2009). What’s
personality to performance motivation: A meta- the evidence on evidence-based management?
analytic review. Journal of Applied Psychology, Academy of Management Perspectives, 23, 5–18.
87, 797–807. Rousseau, D. M. (2006). Is there such a thing
Kluger, A. N., & DeNisi, A. (1996). The effect of feed- as evidence-based management? Academy of
back intervention on performance: A historical Management Review, 31, 256–269.
review, a meta-analysis, and a preliminary feed- Rousseau, D. M., Manning, J., & Denyer, D. (2008).
back intervention theory. Psychological Bulletin, Evidence in management and organizational
119, 254–284. science: Assembling the field’s full weight of
Landy, F. J. (2008). Stereotypes, bias, and personnel scientific knowledge through reflective reviews.
decisions: Strange and stranger. Industrial and Annals of the Academy of Management, 2,
Organizational Psychology: Perspectives on Sci- 475–515.
ence and Practice, 1, 379–392. Rousseau, D. M., & McCarthy, S. (2007). Evidence-
Lawler, E. E., & Mohrman, S. A. (in press). Doing based management: Educating managers from an
research that is useful for theory and practice–25 evidence-based perspective. Academy of Manage-
years later (2nd ed.). San Francisco: Berrett Koehler. ment Learning and Education, 6, 94–101.
22 R.B. Briner and D.M. Rousseau

Rynes, S. L., Brown, K. G., & Colbert, A. E. (2002). Soukup, S. M. (2000). The center for advanced nursing
Seven common misconceptions about human practice evidence-based practice model. Nursing
resource practices: Research findings versus practi- Clinics of North America, 35, 301–309.
tioner beliefs. Academy of Management Executive, Tranfield, D., Denyer, D., & Smart, P. (2003). Toward
16, 92–103. a methodology for developing evidence-informed
Sackett, D. L., Richardson, W. S., Rosenburg, W., & management knowledge by means of system-
Haynes, R. B. (1997). Evidence-based medicine: atic review. British Journal of Management, 14,
How to practice and teach EBM. London: Churchill 207–222.
Livingstone. VanLehn, K., Siler, S., Murray, C., Yamauchi, T., &
Schmidt, F. L. (2009). Select on intelligence. In Bagget, W. B. (2003). Why do only some events
E. A. Locke (Ed.), Handbook of principles of orga- cause learning during human tutoring? Cognition
nizational behavior: Indispensable knowledge for and Instruction, 21, 209–249.
evidence-based management (2nd ed., pp. 3–18). Van de Ven, A. H. (2007). Engaged scholarship: A
New York: Wiley. guide for organizational and social research. New
Silzer, R., Cober, R., Erickson, A., & Robinson, G. York: Oxford University Press.
(2008). Practitioner needs survey: Final sur- Van de Ven, A. H., & Johnson, P. E. (2006). Knowledge
vey report. SIOP Professional Practice Commit- for theory and practice. Academy of Management
tee. Retrieved from www.siop.org/Practitioner%20 Review, 31, 802–821.
Needs%20Survey.pdf
Smith, R. (1991). Where is the wisdom . . .? The poverty
of medical evidence. British Medical Journal, 303,
798–799.

You might also like