You are on page 1of 20

The Science Behind the

Predictive Index Behavioral AssessmentTM

Client Summary Report


May 2020

*This document is intended to provide a simplified and concise overview into the science behind The Predictive Index
Behavioral Assessment. Therefore, this document does not contain all of the scientific details related to the
development, scoring, and statistical properties of the assessment. For a complete and thorough overview, please
see The PI Behavioral Assessment Technical Manual or EFPA report.
The Science Behind the
PI Behavioral Assessment

Table of Contents
About the Predictive Index ............................................................................................................................ 3
The PI Behavioral Assessment ..................................................................................................................... 4
Designed for the Workplace .......................................................................................................................... 5
Proven Predictive Power in the Field ............................................................................................................ 6
Built for the Entire Employee Lifecycle ......................................................................................................... 7
Engineered with Modern Psychometrics ....................................................................................................... 7
Assessment Concepts and Theory ........................................................................................................... 8
Assessment Development, Review, and Translation ............................................................................... 9
Norm Sample .......................................................................................................................................... 10
Validity ..................................................................................................................................................... 10
Reliability ................................................................................................................................................. 12
Fairness................................................................................................................................................... 14
Conclusions ................................................................................................................................................. 18
References .................................................................................................................................................. 19

© 2020 Predictive Index, LLC 2


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

About the Predictive Index


In an era of easy access to technology, information, and customers, talent stands as one of the final
differentiators on which businesses compete. To gain the competitive edge, companies must not just attract
and retain talent, but also develop and tune their employees for the mutual benefit of both the individuals
and the company’s strategy. The alignment of business strategies and talent strategies is known as talent
optimization. Talent optimization empirically aligns strategy and people practices, weaving talent
improvement practices into the everyday workings of a company to nurture an employee workforce that is
specifically calibrated to the company’s strategic objectives.
The Predictive Index (PI) is the leader in talent optimization applications, resources, and training. Serving
more than 8,000 clients across 142 countries and delivering solutions in 70 languages, PI offers a proven
talent optimization methodology that allows businesses to understand the factors driving their workforce so
that they can be aligned to the strategic goals of the company. Through a unique blend of scientific
assessments, ground-breaking software, highly rated management training, and professional consulting
from some of the world’s best workplace behavior experts, PI helps clients overcome their most complex
business challenges. Scientific validation and a 60-year proven track record have shown that business
challenges big and small are no match for PI’s unique approach to client education and knowledge transfer,
which ensures swift adoption, direct return on investment, and high impact on performance metrics.
PI’s flagship assessment, the PI Behavioral Assessment, is one of the cornerstones of PI’s talent
optimization system. Each year, 2.5 million PI Behavioral Assessments are administered and over 7,000
professionals are trained in PI workshops. The PI Behavioral Assessment focuses on work-related
personality scales instead of broad measures of personality, maintaining focus on the behavioral drives that
are relevant to job performance, employee development, succession planning, and change management.
The PI Behavioral Assessment uses a simple, efficient, and approachable checklist format that most
respondents complete in about six minutes, making it easy to administer to large groups of candidates or
employees with minimal disruption. This user-friendly assessment experience belies the state-of-the art
psychometrics and research behind the instrument that ensures clients receive accurate, valid, and
actionable insights about the motivating drives of their workforce. The science of PI’s assessments and user
materials has been developed in a manner consistent with all critical industry standards and guidelines.
These standards are designed to provide a framework for determining the proper use of assessments and
other selection procedures, as well as to prevent discriminatory employment practices.
This document summarizes the key science, research, and psychometric properties of the PI Behavioral
Assessment. Full technical manuals and other research papers from PI are available to clients through their
PI partner. PI’s research portfolio is constantly growing, and clients are encouraged to speak with their PI
Partners for the latest updates from PI’s research team. Table 1 provides a high-level overview of the
psychometric properties of the PI Behavioral Assessment as of May 2020.

Table 1. Summary of Psychometric Properties of the PI Behavioral Assessment (May 2020).


Validity Reliability
Mean Mean
Type Total n Type Total n
Coefficient Coefficient
Criterion Validity 25,000+ 0.30 Cronbach’s Alpha 9,645 0.83
Convergent Validity 1,823 0.43 Test-Retest (0-8 years) 11,684 0.60
Construct Validity 9,645 0.81 Coefficient H 9,645 0.86
Report Accuracy 3,999 0.87
Effectiveness for Employee 691 0.81
Development

© 2020 Predictive Index, LLC 3


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

The PI Behavioral Assessment


The PI Behavioral Assessment is a theory-based self-report measure of normal adult work-related
personality and has been developed and validated extensively and exclusively for use within occupational
and organizational populations. The PI Behavioral Assessment is untimed, takes approximately six minutes
to complete, and is administered through an easy-to-use online format. The assessment employs a free-
choice response format, in which individuals are presented with two questions, each followed by a listing of
adjectives. The first question asks respondents to endorse adjectives which they feel describe the way they
are expected to act by others (the Self-Concept domain). The second question asks respondents to endorse
adjectives which they feel really describe them (the Self domain). Averaging across these two domains
yields a third, implied domain (the Synthesis), which can be interpreted as reflecting an employee’s likely
observed behavior in the workplace.
Each of the adjectives on the PI Behavioral Assessment is associated with one of five factors: Dominance,
Extraversion, Patience, Formality, and Objectivity (referred to as Factors A, B, C, D, and E respectively; see
Table 2). Dominance, Extraversion, Patience, and Formality are primary factors that determine the
respondents’ preferred behavioral drives in the workplace. Objectivity is used as a secondary, modifying
factor describing a respondent’s decision-making style in the expression of the four primary factors. A
respondent’s scores are reported in the Self, Self-Concept and Synthesis domains, and these scores are
presented graphically as behavioral patterns. The patterns are then interpreted by practitioners trained
through a PI workshop, who utilize the PI Behavioral Assessment data for the selection, coaching,
motivation, and development of employees.

Table 2. Five Factors Reported in the PI Behavioral Assessment.


PI Factor Description

The degree to which an individual seeks to control his or her environment.


Individuals who score high on this dimension tend to be independent, assertive
A DOMINANCE
and self-confident. Individuals who score low on this dimension tend to be
agreeable, cooperative and accommodating.

The degree to which an individual seeks social interaction with other people.
Individuals who score high on this dimension tend to be outgoing, persuasive
B EXTRAVERSION
and socially poised. Individuals who score low on this dimension tend to be
serious, introspective and task oriented.

The degree to which an individual seeks consistency and stability in his or her
environment. Individuals who score high on this dimension tend to be patient,
C PATIENCE
consistent and deliberate. Individuals who score low on this dimension tend to
be fast-paced, urgent and intense.

The degree to which an individual seeks to conform to formal rules and


structure. Individuals who score high on this dimension tend to be organized,
D FORMALITY
precise and self-disciplined. Individuals who score low on this dimension tend to
be informal, casual and uninhibited.
The degree to which an individual relies on objectivity when processing
information and making decisions. Individuals who score high on this dimension
E OBJECTIVITY tend to be rational, logical and are typically influenced by facts and data.
Individuals who score low on this dimension tend to be subjective, intuitive and
are typically influenced by feelings and emotions.

© 2020 Predictive Index, LLC 4


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

PI Behavioral Assessment results are reported within-person, meaning that the focus is not on how one’s
personality compares to other people, but instead on whether or not one’s behavioral drives are aligned with
the behavioral demands of the work environment. This approach to assessment maintains comparable
validity and reliability to more conventional personality assessments but has the added benefit of focusing
on how a person is predicted to behave rather than how they compare to other people (e.g., Saville &
Willson, 1991). For example, in the charts below, the candidate’s behavioral pattern is shown first. This
person tends to approach their work with Dominance behaviors (Factor A) followed by Extraversion
behaviors (Factor B) and tends not to express Patience or Formality behaviors (Factors C and D) as
consistently. The second chart shows the demands of their particular job role, as measured with the PI Job
Assessment. This person is an excellent behavioral fit for their role.

Figure 1. PI Behavioral Assessment plots of a candidate (above) showing relative expression of their
behavioral drives. This is compared to a behavioral job target (below) from the PI Job Assessment. This
person is an excellent behavioral fit for their job.

Designed for the Workplace


Almost all jobs have behaviorally driven tasks and hiring and coaching to the ideal behaviors results in
increased performance; however, one must first identify and measure relevant aspects of people’s
personality as it relates to their job. To put this into context, the U.S. Department of Labor measures the
importance of personal characteristics and behaviors for over 950 jobs; i.e., most jobs have some
behavioral component (National Center for O*NET Development, 2020). At the same time, there are a
variety of personality models, and hundreds of personality scales. For example, the International Personality
Item Pool (IPIP) lists more than 250 personality scales (Goldberg et al., 2006). Companies that wish to
leverage measures of personality for workplace use must select personality scales that are:

• Relevant to workplace behavior or performance.


• Valid, reliable, and fair for selection and development in the workplace.
• Comprehensive enough that the scales could be used for all roles at a company.
For example, one would not want to use all the scales of the ubiquitous five-factor model because several of
the scales have little relevance to workplace behavior (e.g., Chen, 2018). For workplace applications, there
is no need to try to map the entire personality space—one needs only to focus on traits that relate to
workplace behaviors and which can inform talent optimization decisions.

© 2020 Predictive Index, LLC 5


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Proven Predictive Power in the Field


Since 1955, the PI Behavioral Assessment has repeatedly met the criteria specified previously: it is relevant
to workplace behavior, it is valid and reliable, and it is general enough that it can be used for any job role. Its
relevance to workplace performance has been demonstrated in hundreds of criterion validity studies, with
significant results found in 94% of nearly 350 studies conducted since 1992. The average criterion validity
coefficient in the significant tests was r = 0.30. These studies were conducted across 111 unique job roles in
11 different industries using assessment scores and performance data from more than 25,000 working
adults, demonstrating the flexibility of the instrument for a variety of roles (Foster et al., 2020).

31%

26%

19%

14%

6%
2% 2%
0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90
Criterion Validity Coefficient (Correlation)

Figure 2. Distribution of criterion validity coefficient effects showing the strength of statistically significant
relationships between PI Behavioral Assessment scores and 4,800 measures of job performance in nearly
350 studies with PI clients. The PI Behavioral Assessment predicted performance in 94% of these studies.
PI continues to expand its portfolio of validity research and the sophistication of methods used in the
analyses. PI clients have worked with PI’s researchers to go beyond correlations to design customized
studies for clients examining how behavior interacts with variables such as tenure or regional differences.

$1,500,000 $1,500,000

$1,000,000 $1,000,000

$500,000 $500,000
■High Dominance = 2.0 σ
■Low Dominance = -2.0 σ
$0 $0
KENNESAW ALPHARETTA WESTMINSTER -3σ -2σ -1σ 0σ 1σ 2σ 3σ
Figure 3. Example of 2017 PI study showing relationship between Dominance behaviors and sales,
accounting for location effects in a sample of U.S. retail salespeople.

© 2020 Predictive Index, LLC 6


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Built for the Entire Employee Lifecycle


AERA, APA, and NCME (2014) identified personal awareness, growth, and action as some of the many
possible uses of psychological assessments, and this describes the PI Behavioral Assessment well. The
assessment results create awareness of behavioral drives and how they manifest as needs or preferences
in a work environment. From this, growth can be achieved by helping employees use that awareness of how
their behavior may help or hinder their performance in the workplace. This awareness of behavioral drives
can also be turned into action for employers matching behavioral drives to specific job roles, facilitating
communication, managing performance, and rewarding employees (The Predictive Index, 2015):
Hiring – Practitioners consider behavioral drives of a job and compare these to the PI Behavioral
Assessment results of individual candidates. For example, a low Extraversion candidate may struggle in a
position that requires him or her to build relationships with new people on a regular basis.
Development – Practitioners help employees understand how their behavioral drives impact how they
engage with others and their roles in team leadership dynamics. For example, a low Patience manager may
find that his or her tendency to quickly change the team’s assignments with little notice might demotivate
employees who need a steady, planned-out work environment.
Change Management – Practitioners advise leaders managing a change at a company by helping them
understand how their employees’ behavioral drives might impact their readiness for a change, their
reactions to the change, and their desire to be engaged with the change process. For example, a company
implementing changes to procedures and operations may find that their high Dominance employees are
resistant to the changes if they do not feel that they can provide their own input on the changes.
Growth Strategy – Practitioners also use the results of the PI Behavioral Assessment to consider how
behavioral drives and needs play a role in training employees and leveraging their awareness of their
behavioral drives to help them succeed in new roles. For example, a manager with low Formality may have
gained the experience needed to lead a department, but if that department’s work requires a lot of attention
to detail and careful adherence to policies, that new manager may want to work closely with other
employees with higher Formality scores to ensure that he or she does not let things fall through the cracks

Engineered with Modern Psychometrics


PI employs a team of experienced scientists and researchers who develop, maintain, monitor, and
document the PI assessments to ensure that they are valid, reliable, and fair. The PI Behavioral
Assessment has been in widespread commercial use since 1955, and the validity and reliability of the PI
Behavioral Assessment have been investigated by numerous PI researchers and third parties (e.g., Barnett
et al., 2017; Everton, 1999; Everton, 2000; Fossey, 2017; Foster et al., 2015; Foster et al., 2019, Harris,
2006; Harris, Tracy, & Fisher, 2014; Perry & Lavori, 1983; Sjöberg, 2000; Sjöberg, 2003).
The most comprehensive third-party review of the PI Behavioral Assessment was conducted from 2017 to
2018 through the certification company DNV-GL. In that review, the PI Behavioral Assessment was
reviewed by two psychologists and independent auditors who evaluated the assessment against the
guidelines published by the European Federation of Psychologists' Associations (EFPA, 2013). This audit
covered validity, reliability, fairness, development, norms, reporting, supporting documentation and training,
and even pricing and distribution. The PI Behavioral Assessment passed this review and was certified under
the EFPA model in September 2018. PI continues to maintain this certification through periodic reviews.
The following sections provide summaries of the technical properties and psychometrics of the PI
Behavioral Assessment; however, PI’s researchers are always adding new analyses and findings. Full
technical reports and white papers are available from PI Partners.

© 2020 Predictive Index, LLC 7


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Assessment Concepts and Theory


When first designing the PI Behavioral Assessment’s primary factors, Daniels (1955) was influenced by
Marston’s (1928) earlier theory of four primary traits. According to Marston’s theory, individuals behave
along two axes, with their actions tending to be either active or passive (axis 1) and their perception of the
environment tending to be either antagonistic or friendly (axis 2). Daniels’ (1955) primary factors for the PI
Behavioral Assessment echoed Marston’s concepts by defining the primary factors as interactions between
environment response (proactive or reactive) and perception of the environment (favorable or unfavorable).
As a result, the PI Behavioral Assessment’s primary factors are described in terms of the motivating needs
and behaviors associated with each factor. The unmet motivating needs represent the driving force behind
the behaviors associated with each factor. This is similar to Marston’s (1928) theory, where he theorizes that
primary emotions (and the behaviors associated with them) are responses to unmet needs (disequilibrium).
Daniels (1955) described the primary factors on the PI Behavioral Assessment as motivating needs and
behaviors to help frame personality in terms that would be applicable to the workplace. Additionally, to help
address workforce applications, Daniels added a secondary factor (Objectivity) to describe how a
respondent processes information and makes decisions. Objectivity is based on Fiske’s (1949) proposed
trait called “recurrent factor: inquiring intellect,” which Fiske derived from aspects of Cattell’s (1946) primary
traits.
The purpose of having two separate sections (Self and Self-Concept) on the PI Behavioral Assessment
checklist is to assess an individual’s response and his or her adjustment to the two different sets of
environmental stimuli to which they feel that they have been and continue to be exposed: the individual’s
own needs and the perceived demands of his or her environment (Daniels, 1955). This important duality
between one’s self and one’s social environment was clearly explicated by Lecky (1945) as part of his self-
consistency theory of personality, as well as by Cantril (1950). MacKinnon (1944) had a similar theory,
suggesting that personality has two conceptually distinct definitions. MacKinnon argued that, when used one
way, personality refers to the distinctive and unique impression that one makes on others. This perspective
refers to personality from the viewpoint of the observer and is functionally equivalent to a person’s
reputation. Used in a second way, personality refers to the structures inside of a person that are useful in
explaining why a person creates a particular impression on others.
Maslow (1943) and Rogers (1959) were two of the first psychologists to establish the notion of self-concept
that is reflected by the Self-Concept domain on the PI Behavioral Assessment. According to Rogers,
everyone strives to reach an “ideal self,” and he hypothesized that psychologically healthy people actively
move away from roles created by others' expectations, and instead look within themselves for validation.
It is these lines of theory that form the basis of asking respondents to describe themselves in both the Self
and Self-Concept domains. Users can review how respondents see themselves and compare this to how
they perceive the demands of their environment. Factor score differences between the Self and the Self-
Concept domains may provide an indication of whether or not one sees goals of value in his or her social-
work environment and how much behavioral adaptation is needed between the respondent’s internal
behavioral drives and the perceived demands of the environment. Mirroring Roger’s (1959) hypothesis of
ideal self, Daniels argued that workers will be more comfortable when there is general alignment between
their Self and Self-Concept scores, but their combined Synthesis pattern is most likely to reflect their
observed behavior.
The use of adjectives as symbols that tap into one’s largely unconscious images of self and self-concept is a
key projective element of the PI Behavioral Assessment. Daniels (1955) cited Kuder (1954) who observed
that open-ended prompts can yield responses that characterize the respondent’s attitude. For example, on
the PI Behavioral Assessment’s Self-Concept checklist, respondents are asked to select adjectives that
reflect how they are perceived by others, but the respondent is left to make his or her own interpretation

© 2020 Predictive Index, LLC 8


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

about who those others may be (e.g., coworkers, clients, friends, family, managers, neighbors). Thus, the
use of unspecific stems and solitary adjectives supports an unconscious projective response that reflects
the respondent’s own perceptions of the major drivers of his or her self-concept.
Finally, it is worth noting that the decision to use the free-choice (rather than the forced-choice) selection of
adjectives in the PI Behavioral Assessment was made with the intent of enhancing the projective value of
the instrument, since a free-choice form provides less direction for the subject than does a forced-choice
form. Daniels argued that by allowing the respondent this flexibility, it better represents one’s stimulus to the
environment. Daniels referenced the work of Rogers (1951), explaining that this format allows one to record
the respondent’s reactions to his or her perception of the environment. The use of a checklist of adjectives
also mirrors the designs of other free-choice personality assessments. Gough (1983) recounts that he used
125 of the 171 adjectives from Cattell’s (1946) study of personality in the development of a checklist-based
personality assessment in 1949, with additional adjectives being added to the form in 1952. Clarke’s (1956)
assessment (mentioned previously) was also based on Marston’s (1928) theories, and it too was designed
as a free-choice checklist. These checklist assessments are still in use today. In addition to the projective
value of allowing the respondent to select adjectives to describe his or her Self and Self-Concept behavioral
configurations, the PI Behavioral Assessment’s adjective checklist has an element of efficiency to it,
allowing respondents to quickly decide whether or not each of the 86 words describes him or herself
(Daniels, 1955).

Assessment Development, Review, and Translation


In 2014, The Predictive Index, together with Caveon and Drasgow Consulting Group, started work on the
current revision of the PI Behavioral Assessment. This revision was undertaken mainly to update the
content of the instrument because word usage changes over time. In addition, analyses were conducted to
improve the psychometric properties of the PI Behavioral Assessment over the previous revision. The
current form of the assessment became commercially available in 2017. The revision of the PI Behavioral
Assessment was constructed using the following best-practice test construction methodology (e.g.,
Schmeiser & Welch, 2006). The following summarizes the major steps in the development of the instrument.
Evaluation of items – On a language-based instrument, such as the PI Behavioral Assessment, word
usage changes over time. Periodically, the items in the instrument should be reviewed and a decision
should be made to continue using them or to replace them with better-performing items (AERA, APA, &
NCME, 2014). Over time, item evaluation methodologies, tools, and techniques are also generally
improved, and these can be used to improve the test’s ability to measure. Following best practices (e.g.,
Schmeiser & Welch, 2006), both quantitative (i.e., statistics) and qualitative (i.e., expert review) measures
were used in the evaluation of the PI Behavioral Assessment adjectives (Foster et al., 2020).
Selection of items – After field testing the draft adjectives and determining their measurement
characteristics using item analyses and content alignment studies with subject matter experts, the adjectives
that provided the best measurement were selected for form construction (Foster et al., 2020).
Assessment of reliability and dimensionality – In order to ensure the selected items measured the
factors as intended and designed, the new form was analyzed quantitatively. The analysis used statistics to
make inferences regarding the form’s reliability and dimensionality (Foster et al., 2020).
Determination of validity – Once the new form was constructed, convergent and discriminant validity were
examined to investigate whether the new revision’s factor scores correlated with previous revisions and with
other closely related constructs, while also showing lower correlations with conceptually different constructs.
Criterion validity was also analyzed with multiple samples of respondents working in different job roles, and
confirmatory factor analysis was conducted to check the factor structure of the instrument (Foster et al.,
2020).

© 2020 Predictive Index, LLC 9


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Norm Sample
Standardizing scores for the PI Behavioral Assessment and ensuring the generalizability of the
psychometric analyses requires a representative sample. PI’s tools are designed to be used with the global
workforce, so it was critical to recruit an international sample for piloting and calibrating the PI Behavioral
Assessment. The PI Behavioral Assessment Norm is based on a sample of 9,645 working adults from 129
countries, and the assessment is now taken by more than two million people around the world every year.
Of the initial sample’s 7,658 respondents who reported their gender, 52% were male and 48% were female
(see Figure 4). The average age of the 7,336 respondents who reported age in the sample was 40.1 years
and ranged between 18 years and 64 years (see Figure 5). Although the sample included diverse
nationalities, 64% percent of the sample was from the U.S. The primary ethnic composition of these U.S.
respondents was Caucasian (74%), followed by African American or Black (12%), Latino, Latina or Hispanic
(8%), and Asian (2%).

938 905 932 923


845 855
707

48% 52% 336

Male Female Respondent's Age


Figure 4. Norm sample gender distribution. (n = 7,658) Figure 5. Norm sample age distribution (n = 7,336)

Validity
The construct validity of the current revision of the PI Behavioral Assessment was examined using
exploratory factor analysis, confirmatory factor analysis, content alignment ratings, corrected item-total
correlations, differential item functioning (DIF) analysis across groups, and convergent correlations with
other personality scales (Foster et al., 2020). Additional validation has been conducted through criterion
validation studies (discussed previously), report accuracy studies, and analysis of the validity of use cases.
Overall, the expansive portfolio of validity evidence described here underscores that the PI Behavioral
Assessment is a valid and powerful tool for assessing workplace behaviors.
PI researchers began validation of the current revision with content alignment studies. The purpose of
content alignment studies was to quantify subject matter expert opinion about the alignment of existing and
potential new words or items with the PI Behavioral Assessment factors. In the current revision, this involved
the 86 existing adjectives of the previous revision of the PI Behavioral Assessment and 140 new pre-test
adjectives. Fifteen content experts, each with a doctorate in psychology, education, or a related field,
participated in content alignment studies. All experts had extensive experience in assessment, education,
and psychology, particularly as they apply to organizational environments. The content experts were asked
to rate the degree to which each word aligns with each PI Behavioral Assessment Factor on a scale of 0 to
100. This information was used to inform item selection along with other data.
To obtain classical test theory item statistics, the new and existing words were administered to operational
samples of respondents. Researchers collected 136,544 usable cases for analysis of the existing version of
the assessment and approximately 10,000 cases for each of the 140 new pre-test words (Foster et al.,

© 2020 Predictive Index, LLC 10


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

2015). These responses were used to calculate selection rates for adjectives and corrected item-total
correlations, which were used to select items for the final form and demonstrate construct validity.
Exploratory factor analyses (EFA) were conducted to evaluate the unidimensionality of each of the factors
on the PI Behavioral Assessment, which was important for validation, as well as a requirement for the
assumptions underlying internal consistency reliability. Scree plots from the EFA show clear
unidimensionality in all five of the measured traits. Foster et al. (2020) also conducted a confirmatory factor
analysis of the factor structure for the PI Behavioral Assessment, testing the assumption that the observed
responses on the PI Behavioral Assessment represented five latent constructs (Factors A – E). Foster et al.
(2016) found that the model fit well (Root Mean Square Error of Approximation [RMSEA]=0.064,
Standardized Root Mean Square Residual [SRMR]=0.043, Goodness of Fit Index [GFI]=0.93, Comparative
Fit Index [CFI]=0.99). Validity coefficients from the factor loadings ranged from 0.73 to 0.85.
PI also conducted two convergent validity analyses to test whether PI Behavioral Assessment scores
correlated with other similar (but not identical) personality scales. The first study had a sample size of n =
1,023, and the second had a sample size of n = 800. Table 3 shows a sample of these results, and the full
results are reported in Foster et al. (2016). The correlational evidence from these studies leads to the
conclusion that, overall, convergent and divergent validity for the PI Behavioral factor scores was obtained

Table 3. Sample of Convergent Validity of the PI BA Self Raw Scores with IPIP Scales.
IPIP Scale A B C D E
Assertiveness 0.50 0.32 -0.07 0.11 0.20
Domineering 0.31 0.10 -0.20 -0.04 -0.01
Extraversion 0.37 0.59 0.10 0.08 0.09
Sociability 0.33 0.54 0.15 0.13 0.12
Calmness ___ 0.23 0.50 0.23 0.22
Good-Nature 0.17 0.30 0.47 0.18 0.16
Patience 0.12 0.20 0.48 0.23 0.21
Methodicalness 0.15 0.12 0.17 0.43 0.23
Conscientiousness 0.24 0.20 0.13 0.37 0.22
Judgment 0.22 0.13 0.13 0.28 0.32
Independent Minded 0.26 ___ 0.18 0.23 0.41
Note: Boldfaced correlations are hypothesized convergent validity coefficients. Convergent validity is shown
by the boldfaced correlation being stronger and more positive than other correlations within each row.

Finally, there is the validation of the use of the PI Behavioral Assessment. As noted previously, PI has
conducted hundreds of criterion validity studies, the results of which demonstrate that the PI Behavioral
Assessment is both predictive and relevant for hiring decisions; however, PI must also make the case that
the assessment is useful for development of existing employees. The utility of the assessment for this
purpose can be demonstrated in two ways. First, PI provides a software system of reports and content that
supports the development use case. This is evidenced by the availability of Coaching Guides, Relationship
Guides, Personal Development Charts, and Team Workstyles reports. The content and design of these
tools is built to support the employee development use case, and each of these items is backed by careful
user testing and iterative improvements based on user feedback. The second way to evaluate whether the
PI Behavioral Assessment is useful for employee development is the evidence collected from client use. Not
all clients use the PI Behavioral Assessment for employee development, but a survey of 1,127 PI clients in
May 2017 showed that 691 (61%) of these clients use the PI Behavioral Assessment for employee
development. PI also tracks client usage at a more granular level through software, but the overall trend

© 2020 Predictive Index, LLC 11


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

provides strong evidence that the PI Behavioral Assessment is useful for employee development at many
companies.
However, just because it is being used does not necessarily mean that it is having the intended
consequences of improving employee development; i.e., companies can certainly engage in employee
development without the PI Behavioral Assessment. To verify that our tools are having the intended benefits
for clients, PI tracks a variety of feedback measures, both from users and respondents. For example, of the
clients using the PI Behavioral Assessment for employee development in that May 2017 study mentioned
previously, 559 clients (81%) agreed or strongly agreed that PI’s tools have “helped us develop better
employees.” We also track how well the reports resonate with the respondents and users, with recent
tracking showing that 87% of people agree or strongly agree with the interpretive text provided in their
reports (only 4% of respondents disagreed with their results, indicating a very high report accuracy).

Reliability
Reliability refers to the precision or consistency of measurement (Nunnally & Bernstein, 1994). A common
way to estimate reliability is by computing internal consistency reliability. Of the measures of internal
consistency reliability, the one used most often is coefficient alpha, which reflects the extent to which items
on a measure are intercorrelated. PI has conducted reliability analyses at several points over the years or
with region-specific sample, but the best representation of the reliability of the current revision of the PI
Behavioral Assessment was calculated with 9,645 results from the norm group sample. Table 4 shows the
reliability estimates (alpha) and the standard error of measurement (SEM) for each factor score.

Table 4. Scale Factor Score Reliability Estimates (Alpha) and Standard Error of Measurement (SEM).
Self Self-Concept Synthesis
Alpha SEM(σ) Alpha SEM(σ) Alpha SEM(σ)
A DOMINANCE 0.81 0.44σ 0.77 0.49σ 0.85 0.33σ
B EXTRAVERSION 0.79 0.41σ 0.79 0.42σ 0.85 0.30σ
C PATIENCE 0.79 0.45σ 0.77 0.48σ 0.83 0.36σ
D FORMALITY 0.85 0.38σ 0.84 0.41σ 0.88 0.29σ
E OBJECTIVITY 0.87 0.44σ 0.84 0.50σ 0.90 0.35σ

In addition to the internal consistency reliability estimates in Table 4, one can consider the construct
reliability estimates, which indicate whether the constructs themselves are expected to be stable and
replicable. The construct reliability can be estimated with Coefficient H, which is a useful construct reliability
estimate; because it is not affected by a loading’s sign, it does not decrease when additional indicators are
added, and it cannot be smaller than the smallest loading in the CFA model (Hancock & Mueller, 2001).
Table 5 reports the Coefficient H values from the CFA analysis detailed in Foster et al. (2020).

Table 5. Construct Reliability Estimates.


Coefficient H
A DOMINANCE 0.87
B EXTRAVERSION 0.87
C PATIENCE 0.86
D FORMALITY 0.84
E OBJECTIVITY 0.85
Note: Norm group sample (n = 9,645)

© 2020 Predictive Index, LLC 12


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Reliability can also refer to stability over time. For example, if the PI Behavioral Assessment is used to
inform a hiring decision, will the information gleaned at the hiring stage still hold true a year later? What
about five years later? The most straightforward approach to estimating this kind of reliability is via repeated
measurements of the same person. The correlation between temporally separated scores is known as the
test-retest reliability or coefficient of stability. Three studies by Everton (1999) and Harris et al. (2014) using
the previous release of the PI Behavioral Assessment found 2-week test-retest reliabilities for primary
factors (Self domain) in the 0.71 to 0.84 range. A later study reported by Harris et al. (2014) investigated
test-retest across six months with an average test-retest reliability across factors of 0.71.
These studies show strong test-retest reliability evidence over short periods in experimental conditions, but it
is also worth looking at operational evidence over longer periods of time. Using the previous revision of the
PI Behavioral Assessment, Fossey (2017) analyzed an operational sample of 11,684 respondents who took
the PI Behavioral Assessment twice. These respondents’ first administration occurred between April 2003
and July 2015, and their second administration occurred between August 2006 and July 2015. The median
amount of time that passed between the two administrations was 270 days but ranged from 1 day to over 11
years. Fossey’s (2017) study used an operational sample, meaning that these retests were not initiated as
part of a research study.
To measure the coefficients of stability over different time periods, Fossey (2017) calculated Pearson
correlations for each six-month retest window sample for the PI Behavioral Assessment Self Factors. Figure
6 shows the coefficients of stability in six-month increments up to a maximum eight-year retest window. For
context, the U.S. Bureau of Labor Statistics (2016) reported that the median tenure for U.S. employees is
only 4.2 years. Linear regression analyses showed that time did not have a large effect on the coefficients of
stability except in the case of the Formality scale, which showed a statistically significant decline in
coefficients of stability over the eight years of possible retest windows (F(1,14) = 4.82, p = 0.05). For the
Formality scale, each year that passed between the two administrations of the assessment is expected to
lower the coefficient of stability by only 0.02.

Self Factor Coefficients of Stability over Time


1.0

0.9

0.8
Coefficient of Stability

0.7

0.6

0.5

0.4

0.3

0.2

0.1

0.0
1 Year 2 Years 3 Years 4 Years 5 Years 6 Years 7 Years 8 Years
n = 4755 n = 1791 n = 1294 n = 920 n = 709 n = 551 n = 443 n = 331 n = 262 n = 181 n = 151 n = 108 n = 77 n = 50 n = 35 n = 26

Retest Time Period / Sample Size


Self A Self B Self C Self D
Figure 6. Primary factor scores’ coefficient of stability over two administrations at varying time intervals on
the previous revision of the PI Behavioral Assessment. The chart illustrates that the four Self factors
maintain relatively consistent coefficients of stability for the first six years. There is more variation in the
coefficients of stability in years 6-8, which may be in part due to the elapsed time or due to sampling bias—
sample sizes decreased as the retest window time periods increased.

© 2020 Predictive Index, LLC 13


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

In evaluating these estimates, Fossey (2017) cited the U.S. Department of Labor (1999), which observed
that acceptable coefficients of stability may be below 0.70 for constructs that have the potential to vary
over time. Crocker and Algina (2008) explained that coefficients of stability will also change based on the
time period between retests. As the coefficient of stability decreases, there is no way to know if it
represents fluctuation in the trait or a lack of reliability in the assessment instrument itself. In practice,
test-retest reliability coefficients largely vary depending on the type of assessment and the time period
between retests. Crocker and Algina (2008) observed that personality assessments like the PI Behavioral
Assessment typically have lower coefficients of stability than other types of assessments, such as
aptitude tests, but that personality assessments can still produce relatively stable measures over time.
Fossey (2017) argued that the one-year test-retest reliability range of 0.61-0.65 exhibited by the previous
revision of the PI Behavioral Assessment did not preclude its use for selection decision-making, and he
noted that other personality assessments with similar test-retest reliability estimates have been
recommended for selection use as long as the results are used in conjunction with other evidence (e.g.,
Axford & Hayes, 2014). To summarize, the PI Behavioral Assessment’s factor scores are stable enough
over time to support hiring decisions, when combined with other relevant sources of information, such as
cognitive ability, experience, and education/training.

Fairness
At its foundation, the issue of assessment fairness is an issue of validity. Any workforce assessment should
be fair for all test takers in the intended population, and one step to supporting fairness is to review items to
help avoid test content bias (Schmeiser & Welch, 2006; AERA, APA, & NCME, 2014). The adjectives on the
PI Behavioral Assessment go through a bias and fairness review before being field tested or being used for
scored administrations. Differential item functioning (DIF) analyses are performed regularly to understand
response differences that might affect scores of various demographic groups. These analyses were also
used to inform the selection of items (Foster, et al, 2020). PI also provides translated versions of the PI
Behavioral Assessment and analyses regional differences when feasible to ensure that interpretation is
comparable for users across the globe. The following sections provide an overview of some of the fairness
research and development that has gone into the PI Behavioral Assessment.\
Demographic Groups
In employment settings, large differences between demographic groups’ scores on hiring assessments can
result in lower rates of selection for certain subgroups. When these lower rates occur in the United States, it
is called adverse impact. To date, there is no evidence to indicate that the inclusion of the PI Behavioral
Assessment in a company’s personnel selection system, either in a compensatory or in a “multiple-hurdle”
selection model, results in adverse impact against any protected class. In fact, in 60 years, there has never
been a successful legal challenge involving the PI Behavioral Assessment. In addition, The Predictive Index
has run the following studies to demonstrate that the use of its assessment tool does not result in adverse
impact:
• Wolman (1991) ran a study to determine whether men and women tended to score differently on
the PI Behavioral Assessment, and whether African Americans, Hispanics, and Caucasians tended
to score differently. His analyses showed that neither gender nor race was significantly related to PI
Behavioral Assessment scores.

• Harris (2004) analyzed PI Behavioral Assessment scores to determine whether they produced
adverse impact based on age. The study showed that for all PI Behavioral Assessment Factors,
there was no significant difference between people over age 40 (the protected class) and people
under age 40, confirming similar findings initially obtained by Everton (1998).

© 2020 Predictive Index, LLC 14


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

• In a 2008 banking industry study of 347 employees working in a variety of jobs (e.g., teller, branch
manager, loan officer), gender and race accounted for less than 2% and 3% of the variability,
respectively, in PI Behavioral Assessment factor scores (Harris, Tracy, & Fisher, 2014).
• DIF analysis was conducted to look for statistical bias based on gender, race/ethnicity, and age.
Adjectives with DIF effect sizes of 0.30 or higher were excluded from consideration for the
construction of the PI Behavioral Assessment (Foster et al., 2015).

• Raw score differences were again analyzed for a sample of nearly 620,000 respondents who took
the PI Behavioral Assessment between 2017 and 2020. Differences between gender and age
groups were negligible (Cohen’s D = 0.09 and 0.12, respectively). A one-way ANOVA test of score
differences by race and ethnicity also yielded negligible effects (Eta squared < 0.001) (Foster et al.,
2020).
PI’s software also uses a match score algorithm to objectively rank candidates’ behavioral match to job
targets set with the PI Job Assessment. To gather additional evidence regarding the fairness of match
scores, an adverse impact simulation study was conducted using various match score cut-offs. For
example, if all candidates with a match score of 10 points were hired, and all those with match scores
between 1 and 9 points were rejected, would selection ratios based on demographic groupings result in
adverse impact, as defined by the 4/5ths (or 80%) rule?
Analysis from this study showed that, regardless of the cut-score used or demographic grouping in question,
there was no evidence of adverse impact. Figures 7a-h demonstrate that at each cut-off match score, the
selection ratio remains well above 80%. This study provides evidence that match scores can mitigate or
reduce bias in the hiring process, and the results further strengthen the fairness evidence for the PI
Behavioral Assessment. It is, however, important to note that this adverse impact analysis was hypothetical
and was conducted with very data from hundreds of thousands of respondents applying to a variety of job
roles. This study is presented for illustrative purposes only, and it is not a substitute for hiring organizations
conducting their own adverse impact analyses based upon their hiring practices. Furthermore, The
Predictive Index recommends against using match scores as a true “cut-off”; rather, match scores should be
viewed as a single data point when making selection decisions.
Nevertheless, the evidence from the DIF procedures, simulated adverse impact analysis, and multiple
demographic comparison studies indicates that the PI Behavioral Assessment is age-, gender- and race-
neutral, and PI believes that the inclusion of a well-validated personality assessment such as the PI
Behavioral Assessment in a company’s personnel selection system can support inclusivity and lead to a
more demographically diverse workforce.

© 2020 Predictive Index, LLC 15


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Figure 7a. Selection Ratio by Age (Under 40 – Figure 7b. Selection Ratio by Gender (Male –
40+) Female)
100% 100%

75% 75%

50% 50%

25% 25%

0% 0%
10 9 8 7 6 5 4 3 2 10 9 8 7 6 5 4 3 2
Match Score Cut-Off Match Score Cut-Off

Figure 7c. Selection Ratio by Ethnicity (White – Figure 7d. Selection Ratio by Ethnicity (White –
Non-White) African American)
100% 100%

75% 75%

50% 50%

25% 25%

0% 0%
10 9 8 7 6 5 4 3 2 10 9 8 7 6 5 4 3 2
Match Score Cut-Off Match Score Cut-Off

Figure 7e. Selection Ratio by Ethnicity (White – Figure 7f. Selection Ratio by Ethnicity (White –
Asian) Latinx or Hispanic)
100% 100%

75% 75%

50% 50%

25% 25%

0% 0%
10 9 8 7 6 5 4 3 2 10 9 8 7 6 5 4 3 2
Match Score Cut-Off Match Score Cut-Off

Figure 7g. Selection Ratio by Ethnicity (White – Figure 7h. Selection Ratio by Ethnicity (White –
Native Hawaiian or Other Pacific Islander) Native American or Alaskan Native)
100% 100%

75% 75%

50% 50%

25% 25%

0% 0%
10 9 8 7 6 5 4 3 2 10 9 8 7 6 5 4 3 2
Match Score Cut-Off Match Score Cut-Off

Selection Ratio 4/5 Cut-Off for Adverse Impact (U.S)

© 2020 Predictive Index, LLC 16


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Reasonable Accommodations
The PI Behavioral Assessment can be used with most disabled populations without any modification or
accommodation. The PI Behavioral Assessment is an untimed assessment that uses an adjective checklist
format, and the respondent can control many aspects of the delivery mode and context. Nevertheless, test
users should still be prepared to provide a reasonable accommodation in appropriate circumstances. In
cases where a disability might prevent the respondent from understanding the content of the assessment,
use of the PI Behavioral Assessment is not recommended.
Reasonable accommodations address the specific disability and its potential interaction with aspects of the
assessment design and administration. At a very high level, one can consider three categories of
disabilities: physical or sensory disabilities, psychological disabilities, and cognitive disabilities. One can also
consider their interactions with three aspects of the PI Behavioral Assessment: context, content, and
response (AERA, APA, NCME, 2014). Context refers to the conditions under which the assessment is
administered, as well as aspects of the assessment design, such as instructions and language format.
Fortunately, in the PI Behavioral Assessment, the respondent has control over many aspects of the context,
so they can take the PI Behavioral Assessment in a context that is most appropriate for them, without ever
requesting accommodations. For example, common context accommodations for other assessments may
include things like extended time, private testing rooms, extra resources, or special equipment like screen
magnifiers, but the PI Behavioral Assessment can be taken by the respondent without an administrator
present, and there are no time limits. As such, the test user does not need to make any changes for the
administration—the respondent has control over much of the format, and these changes will not impact the
validity of the results.
The content of the PI Behavioral Assessment (including its translations) has gone through a bias review,
and the content is not expected to result in any irrelevant score variance in disabled subpopulations. The
content of the PI Behavioral Assessment is also written at a very basic reading level. The English version of
the assessment is measured to have a reading level appropriate for 8- or 9-year-old children across three
readability scales: Flesch-Kincaid Grade Level, SMOG Index, and Automated Readability Index. As such,
the PI Behavioral Assessment content is expected to be accessible for most of the adult population, even if
they have a disability that impacts reading or verbal comprehension. If, however, the respondent has a
disability which will severely limit their ability to understand the instructions or content of the assessment,
then use of the PI Behavioral Assessment is not recommended.
The content of the PI Behavioral Assessment is also considered appropriate for use with respondents who
have psychological disabilities. The PI Behavioral Assessment is a normal measure of personality,
described with five behavioral factors: Dominance, Extraversion, Patience, Formality, and Objectivity. It has
never been shown to be linked to any psychological disorders. This distinction is made for the PI Behavioral
Assessment because other Five Factor assessments include scales for Emotional Stability (or Neuroticism)
that have been shown to have weak relations to personality-based disorders and anxiety disorders. The PI
Behavioral Assessment does not include an Emotional Stability scale, and the assessment results are
expected to be valid for respondents with similar psychological disabilities.
Regional Equivalency
PI employs a Localization Project Manager (LPM) who oversees localization and translation projects of PI
content for client regions. The LPM worked with the assessment developers to translate the instructions and
adjectives for the assessment. To do this, they employed professional linguists provided by PI’s translation
company and in-country reviewers who were in the PI Partner network and who were native speakers of the
translated language. Back-translation and cultural reviews were also conducted (B. Van Raalte, personal
communication, May 19, 2017).

© 2020 Predictive Index, LLC 17


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Before launching the current revision of the PI Behavioral Assessment, comparison norm tables were also
created using North American respondents and the remaining respondents from other regions. PI
researchers compared the expected means and standard deviations in the two tables as well as the
potential differences in factor scores that could occur using the two norm tables (Foster et al., 2020). The
impacts were considered negligible for the purposes of interpretation. The same analysis was conducted
with regional samples from Sweden and Norway as part of the PI’s EFPA certification, with the same region-
neutral findings (Barnett et al., 2017). Because PI serves a global workforce, and because current regional
analyses have not shown any meaningful differences for interpretation, a single, amalgamated global norm
table is currently used for scoring. Nevertheless, PI is committed to providing updated or localized norms
when needed, and new regional analyses are conducted depending on demand and as samples permit.

Conclusions
This document summarizes the science behind the PI Behavioral Assessment; however, this is just an
overview. Clients have access to PI’s portfolio of research, technical manuals, administrator guides,
researchers’ “point-of-view” articles, and white papers. There is expansive scientific evidence to show that
the instrument is a well-constructed, thoroughly validated assessment that supports workplace decision-
making. The PI Behavioral Assessment is constructed following best practices for test development and
maintenance, and PI researchers continue to bolster the evidence for the instruments validity, reliability, and
fairness for its intended applications.
Even more impressive, and often overlooked, is the fact that the instrument takes an average of only six
minutes to complete, yet yields such strong, accurate results that are stable enough to support decisions
spanning years (like hiring). The instrument resonates well with respondents too, with 96% of respondents
providing positive responses on the accuracy of their PI Behavioral Assessment report. The ease of use and
approachable format of the assessment helps it fit in seamlessly in the hiring process or for use with existing
employees, allowing companies to scale behavioral awareness to their entire company as part of their talent
strategy.
The PI Behavioral Assessment has persisted as an industry staple for 60 years, retaining its foundations
and utility for the global workforce while keeping pace with advances in psychometrics and
industrial/organizational psychology. It is well-built as an efficient, versatile, and work-relevant assessment
that supports a variety of applications throughout the employee lifecycle. Clients who use the PI Behavioral
Assessment can rest assured that they are getting a well-maintained, accurate, and useful measure of
workplace personality that will continue to be supported by research for decades to come.

© 2020 Predictive Index, LLC 18


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

References
American Educational Research Association, American Psychological Association, & National Council on
Measurement in Education (2014). Standards for educational and psychological testing.
Washington, DC: American Educational Research Association.
Axford, S. & Hayes, T. L. (2014). Review of the Hogan Development Survey [Revised]. In J. F. Carlson. K.
F. Geisinger, & J. L. Johnson (Eds.), The Nineteenth Mental Measurements Yearbook. Lincoln, NE:
University of Nebraska Press.
Barnett, G., Fossey, A., Black, L., Mulvey, S., Poepsel, M., Mias, E., LaMark, D., Van Raalte, B.,
Dombrowski, R., Van Voorhis, L., Blair-Lamb, K., Kazi, Z., Holsinger, M., & Hosein, A. (2017).
Report on the PI Behavioral Assessment for EFPA Review [technical report]. Westwood, MA: The
Predictive Index.
Cantril, H. (1950). The why of man’s experience. New York, NY: Macmillan Company.
Cattell, R. B. (1946). The description and measurement of personality. New York, NY: Harcourt, Brace &
World.
Chen, A. (2018, October 10). How accurate are personality tests? Scientific American. Available from
https://www.scientificamerican.com/article/how-accurate-are-personality-tests/.
Clarke, W. (1956). The construction of an industrial selection personality test. The Journal of Psychology:
Interdisciplinary and Applied, 41(2), 379-394.
Crocker, L, & Algina, J. (2008). Introduction to classical and modern test theory. Mason, OH: Cengage
Learning.
Daniels, A. S. (1955). Sources of The Predictive Index [technical report]. Wellesley, MA: The Predictive
Index.
Equal Employment Opportunity Commission, U.S. (1978). Uniform guidelines on employment selection
procedures (EEOC Publication No. 43 FR 38295, 38312, Aug. 25, 1978). Washington, DC:
Government Publishing Office.
Equal Employment Opportunity Commission, U.S. (1981). Age discrimination in employment act (EEOC
Publication No. 46 FR 47726, Sept. 29, 1981). Washington, DC: Government Publishing Office.
European Federation of Psychologists’ Associations (EFPA) (2013). EFPA review model for the description
and evaluation of psychological and educational tests: Test review form and notes for reviewers (v
4.2.6). A. Evers, C. Hagemeister, A. Høstmælingen, P. Lindley, J. Muñiz, & A. Sjöberg (Eds.).
Brussels: EFPA. Available from: http://www.efpa.eu/professional-development/assessment.
Everton, W. (1999). A further investigation of the construct validity of The Predictive Index®. (Validity Report
No. 1999.11.01). Wellesley Hills, MA: Predictive Index, LLC (Formerly Praendex, Inc.)
Everton, W. (2000). An investigation of the psychometrics properties of the French Predictive Index checklist
[technical report]. Wellesley Hills, MA: Predictive Index, LLC (Formerly Praendex, Inc.)
Fiske, D. W. (1949) Consistency of the factorial structures of personality ratings from different sources.
Journal of Abnormal and Social Psychology, 2, 329-344.
Fossey, A. (2017). PI Behavioral Assessment: Test-retest reliability [technical report]. Westwood, MA: The
Predictive Index.
Foster, D., Maynes, D., Miller, J., Chernyshenko, O. S., Mead, A., & Drasgow, F. (2015). Meta-analysis
report of The Predictive Index (Form IV). Wellesley Hills, MA: Predictive Index, LLC
Foster, D., Maynes, D., Miller, J., Chernyshenko, S., Mead, A., Drasgow, F., Poepsel, M., Barnett, G.,
Fossey, A., Siminovsky, A., & Mulvey, S. (2020). The Predictive Index Behavioral Assessment
technical manual [technical report]. Westwood, MA: The Predictive Index.
Foster, D., Maynes, D., Miller, J., Parr, R., & Drasgow, F. (2015). Report for development of norms for Form
V of the Predictive Index [technical report]. Westwood, MA: The Predictive Index.

© 2020 Predictive Index, LLC 19


CONFIDENTIAL
The Science Behind the
PI Behavioral Assessment

Goldberg, L. R., Johnson, J. A., Eber, H. W., Hogan, R., Ashton, M. C., Cloninger, C. R., & Gough, H. C.
(2006). The International Personality Item Pool and the future of public-domain personality
measures. Journal of Research in Personality, 40, 84-96.
Gough, H. G., & Heilbrun, A. B. (1983). The Adjective Check List manual. Sunnyvale, CA: Consulting
Psychologists Press.
Hancock, G. R., & Mueller, R. O. (2001). Rethinking construct reliability within latent variable systems. In R.
Cudeck, S. du Toit, & D. Sörbom (Eds.), Structural Equation Modeling: Present and Future — A
Festschrift
Harris, T. (2006). The impact of gender and race on Predictive Index scores in a large South African Sample
[technical document]. Wellesley Hills, MA: Praendex, Inc.
Harris, T. C., Tracy, A. J., & Fisher, G. G. (2014). Predictive Index technical overview. Wellesley Hills, MA:
Predictive Index, LLC (Formerly Praendex, Inc.)
Harris, T.C. (2004). The impact of age on Predictive Index scores: An update. Wellesley Hills, MA:
Predictive Index, LLC.
Kuder, G. F. (1954). The use of inventories as projective devices. Educational and Psychological
Measurement, 14(2).
Lecky, P. (1945). Self-consistency: A theory of personality. New York, NY: Island Press.
MacKinnon, D. W. (1944). The structure of personality. In J. M. Hunt (Ed.), Personality and the Behavior
Disorders (pp. 3-48). New York, NY: Ronald Press.
Marston, W. M. (1928). The emotions of normal people. New York, NY: Harcourt, Brace and Company.
Maslow, A. H. (1943). A theory of human motivation. Psychological Review, 50(4), 370-396.
National Center for O*NET Development. (2020). Work Styles. O*NET OnLine. Retrieved May 15, 2020,
from https://www.onetonline.org/find/descriptor/browse/Work_Styles/
Nunnally, J. C. (1978). Psychometric theory (2nd ed.). New York, NY: McGraw-Hill.
Perry, J. C., & Lavori, P. W. (1983). The Predictive Index: A report on reliability and construct validity
[technical report]. Wellesley Hills, MA: The Predictive Index.
Rogers, C. (1959). A theory of therapy, personality and interpersonal relationships as developed in the
client-centered framework. In S. Koch (Ed.), Psychology: A Study of a Science. Vol. 3: Formulations
of the Person and the Social Context (pp. 184-256). New York, NY: McGraw Hill.
Rogers, C. R. (1951). Client centered therapy. Boston, MA: Houghton Mifflin Co.
Saville, P., & Willson, E. (1991). The reliability and validity of normative and ipsative approaches in the
measurement of personality. Journal of Occupational Psychology, 64, 219-238.
Schmeiser, C. B., & Welch, C. J. 2006). Test development. In R. L. Brennan (Ed.), Educational
Sjöberg, L. (2000). The psychometric structure of the Swedish version of the Predictive Index (PI).
Stockholm, Sweden: Stockholm School of Economics.
Sjöberg, L. (2003). Properties of the new Swedish version of the Predictive Index. Stockholm, Sweden:
Stockholm School of Economics.
The International Test Commission. (2006). International guidelines on computer-based and internet
delivered testing. International Journal of Testing, 6(2), 143-171.
The Predictive Index. (2015). Predictive Index management workshop [user document]. Westwood, MA:
The Predictive Index.
U.S. Bureau of Labor Statistics (2016). Employee tenure summary. Washington, DC: U.S. Bureau of Labor
Statistics. Available from https://www.bls.gov/news.release/tenure.nr0.htm.
U.S. Department of Labor (1999). Testing and assessment: An employer’s guide to good practices.
Washington, DC: U.S. Department of Labor. Available from:
https://wdr.doleta.gov/opr/FULLTEXT/99-testassess.pdf.
Wolman, R. N. (1991). Predictive Index factors in relation to race, sex, and job levels. Wellesley Hills, MA:
Predictive Index, LLC (Formerly Praendex, Inc.).

© 2020 Predictive Index, LLC 20


CONFIDENTIAL

You might also like