Professional Documents
Culture Documents
Objectives
After reading this chapter, you should be able to:
1. Explain why evaluation is important.
2. Identify and choose outcomes to evaluate a training program.
3. Discuss the process used to plan and implement a good training evaluation.
4. Discuss the strengths and weaknesses of different evaluation designs.
5. Choose the appropriate evaluation design based on the characteristics of the company and the importance and
purpose of the training.
6. Conduct a cost-benefit analysis for a training program.
INTRODUCTION
Training function was interested in assessing the effectiveness of training programs
Training effectiveness refers to the benefits that the company and the trainees receive from training.
Benefits for trainees may include learning new skills or behavior.
Benefits for the company may include increased sales and more satisfied customers.
Training evaluation measures specific outcomes or criteria to determine the benefits of the program. It refers to the
process of collecting the outcomes needed to determine whether training is effective.
Training outcomes or criteria refer to measures that the trainer and the company use to evaluate training programs.
The evaluation design refers to the collection of information (including what, when, how, and from whom) that will
be used to determine the effectiveness of the training program.
The influence of training is largest for organizational performance outcomes and human resource outcomes and
weakest for financial outcomes
TRAINING EVALUATION
Provides a way to understand the investments that training produces and provides information needed to
improve training
Provides the data needed to demonstrate that training does offer benefits to the company.
Involves both formative and summative evaluation.
FORMATIVE EVALUATION
Refers to the evaluation of training that takes place during program design and development. And helps to
ensure that :
(1) the training program is well organized and runs smoothly
(2) trainees learn and are satisfied with the program.
Usually involves collecting qualitative data about the program.
Qualitative data include opinions, beliefs, and feelings about the program.
Ask customers, employees, managers, and subject-matter experts their opinions on the training content
,objectives ,the program design and to evaluate the clarity and ease of use of a part of the training program
that is provided to them in the way that it will be delivered (ex. online, face-to-face, video)
It is conducted either individually or in groups. Trainers may also be involved to measure the time requirements
of the program.
As a result, training content may be changed to be more accurate, easier to understand, or more appealing.
Pilot testing refers to the process of previewing the training program with potential trainees and managers or with
other customers.
Can be used as a “dress rehearsal” to show the program to managers, trainees, and customers.
It should also be used for formative evaluation.
SUMMATIVE EVALUATION
refers to an evaluation conducted to determine the extent to which trainees have changed as a result of
participating in the training program.
Include measuring the monetary benefits (also known as return on investment) that the company receives from
the program.
Usually involves collecting quantitative (numerical) data through tests, ratings of behavior, or objective
measures of performance such as volume of sales, accidents, or patents.
1. To identify the program’s strengths and weaknesses. This includes determining if the program is meeting the
learning objectives, if the quality of the learning environment is satisfactory, and if transfer of training to the job is
occurring.
2. To assess whether the content, organization, and administration of the program— including the schedule,
accommodations, trainers, and materials—contribute to learning and the use of training content on the job.
3. To identify which trainees benefit most or least from the program.
4. To assist in marketing programs through the collection of information from participants about whether they would
recommend the program to others, why they attended the program, and their level of satisfaction with the program.
5. To determine the financial benefits and costs of the program.
6. To compare the costs and benefits of training versus non training investments (such as work redesign or a better
employee selection system).
7. To compare the costs and benefits of different training programs to choose the best
program.
The hierarchical nature of Kirkpatrick’s framework suggests that higher level outcomes (result) should not be
measured unless positive changes occur in lower level outcomes (learning).
The framework has been criticized for a number of reasons.
First, research has not found that each level is caused by the level that precedes it in the framework, nor does
evidence suggest that the levels differ in importance.
Second, the approach does not take into account the purpose of the evaluation. The outcomes used for evaluation
should relate to the training needs, the program learning objectives, and the strategic reasons for training.
Third, use of the approach suggests that outcomes can and should be collected in an orderly manner, that is,
measures of reaction followed by measures of learning, behavior, and results. Realistically, learning measures
need to be collected at approximately the same time as reaction measures, near the end of the training program, in
order to determine whether learning has occurred.
As a result of these criticisms, additional training outcomes have been added to Kirkpatrick’s original framework .
Training outcomes have been classified into six categories: reaction outcomes, learning or cognitive outcomes,
behavior and skill-based outcomes, affective outcomes, results, and return on investment.
These criteria also help to determine whether the benefits of training exceed their costs.
TRAINING OUTCOMES
1. REACTION OUTCOMES
Refer to trainees’ perceptions of the program, including the facilities, trainers, and content.
Reaction outcomes are often referred to as a measure of “creature comfort.”
Are often called class or instructor evaluations.
This information is typically collected at the program’s conclusion.
Reactions have been found to have the strongest relationship with posttraining motivation, trainee self-efficacy,
and declarative knowledge when technology is used for instructional delivery. This suggests that for online or e-
learning training methods, it is important to ensure that it is easy for trainees to access them and the training
content is meaningful, i.e., linked to their current job experiences, tasks, or work issues.
Trainees may be asked to provide ratings of their own behavior or skills (self-ratings). Peers, managers, and
subordinates may also be asked to rate trainees’ behavior or skills based on their observations.
Because research suggests that the use of only self-ratings likely results in an inaccurately positive assessment of skill
or behavior transfer of training, it is recommended that skill or behavior ratings be collected from multiple perspectives
(ex. managers and subordinates or peers).
4. AFFECTIVE OUTCOMES
Include attitudes and motivation.
Affective outcomes that might be collected in an evaluation include tolerance for diversity, motivation to learn,
safety attitudes, and customer service orientation. And can be measured using surveys.
If trainees were asked about their attitudes on a survey, that
would be considered a learning measure. For example, attitudes toward career goals and
interests might be an appropriate outcome to use to evaluate training focusing on employees self-managing their
careers.
5. RESULTS
Are used to determine the training program’s payoff for the company.
Examples of results outcomes include increased production and reduced costs related to
employee turnover, accidents, and equipment downtime as well as improvements in
product quality or customer service.
Indirect costs are not related directly to the design, development, or delivery of the training program.
Include general office supplies, facilities, equipment, and related expenses; travel and expenses not
directly billed to one program; training department management and staff salaries not related to any one program; and
administrative and staff support salaries.
Benefits are the value that the company gains from the training program.
CRITERIA RELEVANCE
the extent to which training outcomes are related to the
learned capabilities emphasized in the training program. The outcomes need to be valid measures of learning
RELIABILITY
the degree to which outcomes can be measured consistently over time.
DISCRIMINATION
the degree to which trainees’ performance on the outcome actually reflects true differences in performance.
PRACTICALITY
the ease with which the outcome measures can be collected.
EVALUATION PRACTICES
It is important to recognize the limitations of choosing to measure only reaction and cognitive outcomes.
To ensure an adequate training evaluation, companies must collect outcome measures related to both learning and
transfer.
Note that outcome measures are not perfectly related to each other. Research indicates that the relationships among
reaction, cognitive, behavior, and results outcomes are small
The best training outcomes measure was depends on the training objectives.
There is no accepted standard for when the different training outcomes should be collected.
Positive transfer of training is demonstrated when learning occurs and positive changes in skill-based, affective, or
results outcomes are also observed.
No transfer of training is demonstrated if learning occurs but no changes are observed in skill-based, affective, or
learning outcomes.
Negative transfer of training is evident when learning occurs but skills, affective outcomes, or results are less than at
pretraining levels.
Results of evaluation studies that find no transfer or negative transfer suggest that the trainer and the manager need
to investigate whether a good learning environment (e.g., opportunities for feedback and practice) was provided in the
training program, trainees were motivated and able to learn, and the needs assessment correctly identified training
needs.
EVALUTION DESIGN
Determines the confidence that can be placed in the results or how sure a company can be that training is either
responsible for changes in evaluation outcomes or has failed to influence the outcomes.
No evaluation design can ensure that the results of the evaluator are completely due to training. What the
evaluation strives for is to use the most rigorous design possible to rule out alternative explanations for the
results of the evaluation.
Threats to validity refer to factors that will lead an evaluator to question either the:
Internal Validity or the believability of the study results
External Validity or the extent to which the evaluation results are generalized to other groups of trainees and
situations
RANDOM ASSIGNMENT
refers to assigning employees to the training or comparison group on the basis of chance and often impractical.
It helps to ensure that trainees are similar in individual differences such as age, gender, ability, and
motivation.
It helps to ensures that these characteristics are equally distributed in the comparison group and the training
group.
It helps to reduce the effects of employees dropping out of the study (mortality) and differences between the
training group and comparison group in ability, knowledge, skill, or other personal characteristics.
Posttest-Only Design
An evaluation design in which only posttraining outcomes are collected. This design can be strengthened by
adding a comparison group.
The posttest-only design is appropriate when trainees and the comparison group can be expected to have similar
levels of knowledge, behavior, or results outcomes prior to training.
Pretest/Posttest
An evaluation design in which both pretraining and posttraining outcome measures are collected.
The lack of a comparison group makes it difficult to rule out the effects of business conditions or other factors
as explanations for changes.
This design is often used by companies that want to evaluate a training program but are uncomfortable with
excluding certain employees or that only intend to train a small group of employees.
Statistical procedures
known as analysis of variance and analysis of covariance.
Were used to test for differences between pretraining measures and posttraining measures for each condition and
differences between each of the training conditions and the comparison group were analyzed.
These procedures determine whether differences between the groups are large enough to conclude with a high
degree of confidence that the differences were caused by training.
Time series
refers to an evaluation design in which training outcomes are collected at periodic intervals both before and after
training. (training outcomes are collected only once after and maybe once before training.)
The strength of this design can be improved by using reversal ( refers to a time period in which participants no
longer receive the training intervention)
A comparison group can also be used with a time series design.
advantage of the time series design is that:
It allows an analysis of the stability of training outcomes over time.
It helps to rule out alternative explanations for the evaluation results.
Used to evaluate training programs that focus on improving readily observable outcomes (such as accident rates,
productivity, and absenteeism) that vary over time.
A more rigorous evaluation design (pretest/posttest with comparison group) should be considered if any of the
following conditions are true:
1. The evaluation results can be used to change the program.
2. The training program is ongoing and has the potential to have an important influence on (employees or customers).
3. The training program involves multiple classes and a large number of trainees.
4. Cost justification for training is based on numerical indicators. (Here the company has a strong orientation toward
evaluation.)
5. Trainers or others in the company have the expertise (or the budget to purchase expertise from outside the company)
to design and evaluate the data collected from an evaluation study.
6. The cost of the training creates a need to show that it works.
7. There is sufficient time for conducting an evaluation. Here, information regarding training
effectiveness is not needed immediately.
8. There is interest in measuring change (in knowledge, behavior, skill, etc.) from pretraining levels or in comparing
two or more different programs.
Cost-benefit analysis is the process of determining the economic benefits of a training program using accounting
methods that look at training costs and benefits.
ROI provides evidence of the economic value provided by training and development programs.
The following sections discuss how to determine costs and benefits and provide examples of
cost-benefit analysis and ROI calculations.
ROI should be limited only to certain training programs because ROI analysis can be costly.
ROI analysis is best for training programs that are focused on an operational issue (measurable identifiable
outcomes are available), are linked to a company wide strategy (e.g., better customer service), are expensive,
are highly visible, have management interest, are attended by many employees, and are permanent.
Determining Costs
Resource requirements model compares equipment, facilities, personnel, and materials costs across different stages
of the training process (needs assessment, development, training design, implementation, and evaluation).
It can help determine overall differences in costs among training programs. Also, costs incurred at different stages of
the training process can be compared across programs.
Accounting can also be used to calculate costs.
One-time costs include those related to needs assessment and program development.
Costs per offering relate to training site rental fees, trainer salaries, and other costs that are realized every time the
program is offered.
Costs per trainee include meals, materials, and lost productivity or expenses incurred to replace the trainees while
they attend training.
Determining Benefits
Helpful Methods in identifying the benefits of training:
1. Technical, academic, and practitioner literature summarizes the benefits that have been shown to relate to a
specific training program.
2. Pilot training programs assess the benefits from a small group of trainees before a company commits more
resources.
3. Observance of successful job performers helps a company determine what successful job performers do
differently than unsuccessful job performers.
4. Trainees and their managers provide estimates of training benefits.
Utility analysis a cost-benefit analysis method that involves assessing the cost of training based on:
Estimates of the difference in job performance between trained and untrained employees
The number of individuals trained
The length of time a training program is expected to influence performance
The variability in job performance in the untrained group of employees.
Utility analysis requires the use of a pretest/posttest design with a comparison group to obtain an estimate of the
difference in job performance for trained versus untrained employees.
Other types of economic analyses evaluate training as it benefits the firm or the government using direct and indirect
training costs, government incentives paid for training, wage increases received by trainees as a result of completion of
training, tax rates, and discount rates.
Success cases
Refer to concrete examples of the impact of training
Show how learning has led to results that the company finds worthwhile and the managers find credible.
Success cases do not attempt to isolate the influence of training but rather to provide evidence that it was useful.
Workforce analytics
Practice of using quantitative methods and scientific methods to analyze data from human resource databases and
other databases to influence important company metrics