Professional Documents
Culture Documents
Evaluation
It would simply mean the act of judging whether the activity to be evaluated is worthwhile in
terms of set Criteria
The ‘worth’, which means the Value, Merit or Excellence of the thing
Stake Holders
Participants
Organization
Training Agency
Trainers
Purpose of Evaluation
Benefits of Evaluation
Better discrimination of training activities between those that are worthy of support and those
that should be dropped
Better integration of training offered and on-the job development
Evidence of the contribution that training and development are making to the organization
Definition
blin (1970) -
Any attempt to obtain information (feedback) on the effects of training programme and to
assess the value of training in the light of that information for improving for the training.
Bramley (1966)-
Validation
Internal- Focus is on the immediacy of the training and learning resulting in the improvement of
the trainee.
External- Series of tests designed to ascertain whether the objectives of internally valid
programme are based on an accurate identification of training needs.
Formative Evaluation
Summative Evaluation
Takes place at the end of training. It examines the outcome from training with eye on fresh
training needs and initiatives.
Participants
Training Agency
Trainers
Organization
Identification of gaps
Validity of training
Training Evaluation:
Why to evaluate?
When to evaluate?
What to evaluate?
How to evaluate?
Why Evaluate?
Training department
Trainees
Conclusive feedback- To establish links between training and organizational needs and goals
Directive feedback- To improve quality, design and delivery of the present and future training.
Intervention feedback- Facilitate transfer of training to the job by identifying support systems
required by training utilizers.
Principles of Evaluation
1) Clarity: Purpose of evaluation in order to set the standards and criteria of evaluation.
2) Objectivity:
4) Feasibility:
• Is it cost effective?
Evaluation has to begin before training activity and end much after the conclusion of visible
training activity.
It should guide trainers for the current training effectiveness and to improve subsequent
training.
Generalization drawn from one evaluation may not identify strengths and weaknesses of
training meant for different set of objectives.
Role of Evaluator
It should be based on sound working relationship with those directly and indirectly involved in
training.
1. The trainees
a) Be acceptable to others.
Paper VII
Training Evaluation
Why to evaluate?
When to evaluate?
What to evaluate?
How to evaluate?
MODELS OF EVALUATION
Reaction
Learning
Job Behavior
Functioning
Ultimate Value
Results – What were the tangible results of the programme in terms of reduced cost ,
improved quality etc?
a) Immediate outcome
b) Intermediate outcome
Pre-Training Evaluation
Post-Training Evaluation
i) Reaction Evaluation
ii) Learning
Follow-up of Evaluation
i) Organisational Level
Changes in Behaviour
Changes in Learning
i) Developmental Stage
ii) Pilot Testing
i) Validation
Stages of Evaluation
A) Pre-Training Evaluation
v) Input Evaluation
i) Reaction Evaluation
v) Ultimate Value
Evaluation Methods
A) Pre-Training Evaluation
c) How does the training department promote the process of training need identification?
Needs of the trainee are person specific, performance standards are job specific
Performance standards are identified based on KRAs e.g. productivity, profit, accident rate ,
machine down time etc
To assess whether there is good congruence between the trainee’s needs and training
objectives
some measure of pre-training knowledge, skill & attitude is desirable to compare with
post training performance
v) Content Evaluation
Committee approach has been found to be fairly successful for input evaluation
Brainstorming session
i) Observation
Normally this is done by using semantic differential scale, three point or five point.
i) Reaction Evaluation
At the end of the course, reactions are collected by use of rating scales of five point or seven
point. The words ‘extremely’, ‘totally', 'very well', 'very bad’ are semantic differential, carry different
meanings for different respondents.
a) Knowledge Learning
i) Objective tests
Levels of knowledge
• Matching items
• Writing essays
b) skill Learning
d) Easy to measure technical skill or physical actions than to assess management or social skills.
c) Attitude Learning
Attitude exist only in the mind and cannot directly inferred from people’s behaviour
Common way of evaluating attitudes and opinions is to hand out a questionnaire at the start &
at the end of the programme
It is the crucial half way stage between training & its ultimate effects
Processing the job improvement plan is possible only through content analysis
The trainee’s boss can enable the trainee to translate the improved behaviour into perceptible
measurable benefits by ensuring post training debriefing discussion with special reference to the
trainee's job improvement plan
G) Follow Up Results
This is basically to seek information on the degree of application of learning & job Improvement
plan
Training Evaluation:
Why to evaluate?
When to evaluate?
What to evaluate?
How to evaluate?
Reactions of the trainees are important but organizers shall look for
• Learning of participants vary, based on value they attach to each topics or modules according
their own requirements and perceptions.
• It is difficult to maintain focus if large numbers of Guest speaker are associated.
Outcome- Control
Outcome- feedback
In this approach growth and development of individual is primary and cost incurred on training is
secondary.
Outcome- information
ii)Techniques of Measurement
1) Observation :
It is most direct, objective and reliable form of measurement for understanding human
behavior.
Systematic observation is considered the hallmark strategy for planning and evaluation of
training programmes.
Types of observation
Most natural, uncontrolled observation to most exact film recording of training sessions.
Aids in observation
Strengths-
i Flexible
ii Simple
Limitations –
2) Interview
Involves both verbal and non verbal communication like gestures, facial expressions, glances
and pauses which reveal subtle feelings
Interviews are used to elicit both quantitative and qualitative information on complex issues like
beliefs, attitudes, feelings and behaviors of the trainees.
It can provide a basis for prediction, understanding and action, thereby providing an active tool
for assessment of the ultimate value of training.
A good interview is one that is planned, executed skillfully and goal oriented.
Types of interviews
i. Structured interview
iv. Sincerity
v. Acceptance
vi. Empathy
vii. Good communication skills (verbal and non verbal)
x. Listening
Merits :
Interview is a superior technique for exploitation of deeper and complex areas of training
evaluation.
Demerits:
Skills of evaluation.
Interviewer bias.
3) Questionnaire
By using questionnaire one can document the effect of training and establish individual profiles
and sets of group profiles.
A questionnaire is useful tool for obtaining reliable and valuable information about beliefs,
feelings and attitudes of trainees.
Questionnaire are often used to measure participants’ attitudes about the areas which may be
affected by training
Questions should be properly worded, arranged, codified and duly pre tested and approved by
the experts in the field.
Merits :
Economical in terms of time, effort and cost for the trainer and trainees.
Demerits:
4) Rating scales
Rating scales methods are primarily used for systematizing and structuring the collection of data
of training results.
The great ease with which they can be administered gives them unusual appeal.
Paper and pencil tests may be used to measure work skills, general intelligence, special
achievement and personality..
These tests are more appropriate for assessing specific attributes to be incorporated in the
training package.
They are widely used in personal selection, placement and to predict job performance.
6)Work samples
A work sample is a miniature replica of the job. The taste is actually a part of the work to be
performed in the job.
This can yield numerical information about how well training participants can perform a
particular task.
Work samples are used in both blue and white collar jobs.
7) Simulation
The work sample is a mirror part of the job, while simulation is a microcosm of the entire job.
8) Job performance
Performance measures are used in assessing training needs, providing feedback to the
participants and in training evaluation.
The basic categories of individuals and group performance measurement are quantity, quality
and timeliness.
Desirable behaviors like team work, maintaining good interpersonal relationships etc.
Problems of measurement
The people behaved differently precisely because they were the subject of research.
Group is swayed by a few leaders who emerge during the course of training.
5) Measurement of change.
4: Final Results
Results
3: Change in B.V.
Behavior
2: KSA
Learning
1 : Exp
Reaction
There is no guarantee that this investment will lead to the result we are looking for
Sending all these employees to training will reduce the working force during the period
Return on Investment
Rooted in Manufacturing.
i. Organizational goals
Long term benefits- Improved Human Relations, Better Communication Ability, Teamwork etc
2. Collect Data
5. Calculate ROI
Training Cost
Fixed Cost
Opportunity cost-
Training Costs
Developmental Costs.
Direct Costs.
Indirect Costs.
Participant’s Compensation.
Evaluation Costs.
Training Cost
Development costs
Program materials
Instructor/facilitator costs
Facilities costs
Travel/lodging/meals
Administrative/overhead costs
Evaluation costs
◦ There are costs at every step – make sure to account for them all.
• Questionnaire
• Surveys
• Tests
• Interviews
• Focus groups
• Observation
• Performance records
• Program follow up
• Project assignments
• Questionnaire
• Surveys
• Tests
• Interviews
• Focus groups
• Observation
• Performance records
• Program follow up
• Project assignments
The Four Major Categories of Hard Data
Primary Measurements
of Improvement
“Hard Data”
Objectively Based
• Scrap
• Waste • Operating Costs
• Rejects • Number of Cost Reductions
• Error Rates • Project Cost Savings
• Rework • Accident Costs
• Shortages • Program Costs
• Product Defects • Sales Expense
• Deviation From
Standard
Major Categories of Soft Data
Work
Habits
Typical Measures
of Improvement
“Soft Data”
New
Skills
Characteristics of Soft Data
Behaviorally Oriented
Change in:
◦ Allow time for change in attitude or behavior, then measure these changes and report
qualitatively.
◦ Allow time for change in productivity, then measure for data and report quantitatively.
• Learning
• Business Impact
• Return on Investment
• Intangible Measures
Common Intangible Variables Linked with Training and Performance Improvement Initiatives
Job satisfaction
Organizational commitment
Work climate
Employee complaints
Employee grievances
Employee stress reduction
Employee tenure
Employee absenteeism
Employee turnover
Employee lateness
Innovation
Community image
Investor image
Customer complaints
Customer loyalty
Teamwork
Cooperation
Conflict
Decisiveness
Communication
Models
Benefit/Cost Ratio
Program Benefits
BCR=
Program Costs
Benefit −Cost
ROI (%)= x 100
Cost
40 clerks.
No of weeks-1
Ans- 0.433
Five hours per week were spent correcting errors before training.
40 clerks.
No of weeks- 4
Calculate ROI %
Ans- 73.33%
by Training Training
training investment
Paper VII
Data Analysis & Statistical Methods.
Training Evaluation:
Why to evaluate?
When to evaluate?
What to evaluate?
How to evaluate?
• Statistics is the method of analysising quantitative data obtained on Training Results from
Groups of Trainees.
• The main aim of statistics is to reduce large quantities of data to manageable and
understandable form.
Introduction:
Some Basic concepts
ii) Drawing of inferences about a body of data when only a part of the data is observed.
Role of statisticians
To present and interpret the results to researchers and other decision makers
A population:
It is the largest collection of values of a random variable for which we have an interest at a particular
time.
For example:
Sample:
It is a part of a population.
For example:
Data
We may define data as figures. Figures result from the process of counting or from taking a
measurement.
For example:
Sources of
data
Comprehensive Sample
Sources of Data:
We search for suitable data to serve as the raw material for our investigation.
Such data are available from one or more of the following sources:
For example:
- The data needed to answer a question may already exist in the form of published
reports, commercially available data banks, or the research literature, i.e. someone else has
already asked the same question.
- The data needed to answer a question may already exist in the form of published
reports, commercially available data banks, or the research literature, i.e. someone else has
already asked the same question.
- iii) Surveys:
- The source may be a survey, if the data needed is about answering certain questions.
- For example:
- If the administrator of a Training Institution wishes to obtain information regarding the mode
of transportation used by Trainees to come to Institution, then a survey may be conducted
among
- iv) Experiments.
- result of an Experiment.
- For example:
- If a Training Institution wishes to know which of several strategies is best for maximizing
attendance of Trainees.
- They might conduct an experiment in which the different strategies are tried with
different Training Programme .
- 1. Center: A representative or average value that indicates where the middle of the data set is
located
- 2. Variation: A measure of the amount that the values vary among themselves
- 3. Distribution: The nature or shape of the distribution of data (such as bell-shaped, uniform, or
skewed)
- 4. Outliers: Sample values that lie very far away from the vast majority of other sample values
A variable:
For example:
- heart rate,
Quantitative Qualitative
Quantitati
ve Qualitati
continuou ve
s nominal
Quantitati
ve Qualitati
descrete ve
ordinal
Quantitative Variables
For example:
- Qualitative Variables
- Many characteristics are not capable of being measured. Some of them can be ordered or
ranked.
- For example:
- For example:
A continuous variable
can assume any value within a specified relevant interval of values assumed by the variable.
For example:
- Height,
- weight,
No matter how close together the observed heights of two people, we can find another person whose
height falls somewhere in between.
Mean – Arithmetic mean is the sum of all the scores divided by the number of score.
Median – Median is the midpoint of the distribution of scores. Median is not affected by outliers.
Mode- Mode is the most frequent scores in the distribution. It is the most commonly occurring value.
Measure of Dispersion
• The Range.
• The Variance.
Scales of Measurements –
1) Nominal Scales – The term nominal pertains to the acts of naming. Nominal Scale often called
Categorical scales, are the simple type. e.g. sex, survey response yes/no.
3) Interval Scales – Interval scale provide far more information about observations and can be
mathematically manipulated with for greater confidence & precision than nominal & ordinal Scales.
Intelligence tests are one good example of an interval scale.
Statistics.
Types of Samples –
A. Probability Sampling –
ii. Stratified Sampling – The population is divided into strata, from which random samples are
drawn. e.g. Men and Women.
iii. Cluster Sampling – it is successive random sampling of units, or sets and subsets. This is useful
when the target population is dispersed throughout a large geographic region.
iv. Systematic Sampling – In systematic sampling the first sample element is randomly chosen
from numbers through K & subsequent elements are chosen every k th interval.
B) Non-Probability Sampling-
i. Quota Sampling – In Quota sampling, knowledge of strata of the population e.g. Sex, region etc.
used to select sample members that are representative, typical & suitable for certain purpose.
ii. Purposive Sampling – It is characterised by use of judgement & a deliberate effort to obtain
representative samples by including typical groups of trainees in the samples.
iii. iii. Accidental Sampling - Accidental Sampling is the weakest form of sampling. In this, one
takes available sample at hand.
ii. Measures of relationship or correlation between two sets of data (before – after, experimental-
control)
iii. Graphic methods for displaying differences between two groups over a period of time.
i. Bar Chart – the bar diagram is often used to compare relative amount of same trait like
intelligence or achievement possessed by two or more groups.
Bar Chart
Pie Diagram – The pie diagram is useful to illustrate populations of the total in striking way.
Pie Diagram
Line Graph- A line graph is used to depict age-progress curve or trend lines.
Line Graph
Histogram - Frequency distribution of metric variables can be portrayed effectively in graphs.
Histogram
PaperVII
i)Tests and Evaluation
ii) Report Writing
Test –
Sample of behavior
There are established rules for scanning, or obtaining quantitative information from the
behavior sample.
Types of tests
3. Self Report- Trainee expresses own feelings, beliefs , interests & attitudes.
3) Mode of administration-
1) Item selection-
To determine what domain and what stage of the training evaluation the test responses will
represent.
What a given test is expected to produce. e.g. test for manager include items on decision
making, problem solving skills.
2) Item formats-
Open ended or Restricted- Restricted are forced choice e.g. True/False, Open ended items
allow more projective responses.
Objective or Projective- Objective items include multiple choice. Projective items are
deliberately vague & ambiguous. Interpretive system is must to derive meaningful scoring e.g.
Impact analysis of organisational effectiveness in terms of KRA & TQM
3) Standardization-
a) Standard Administration
b) Standard Scoring
Characteristics of tests
1. Reliability- Reliability Index is the degree to which test yields consistent results on testing &
retesting.
2. Validity- Validity Index refers to degree to which a test measures what it purports to measure.
3. Cross validation – Independent determination of validity of the entire test is known as cross
validation. It is computed on a different samples of persons.
4. Norms-Norm is normal or average performance. No single set of norms can be used with all
types of trainees & Training Groups.
Objectivity, clarity of direction, ease of application, convenience of scoring, adequacy of norms, cost,
etc.
4. Reputation of department
EVALUATION AUDIT
1. Are procedures established for continual evaluation and quality control of training even if you
are not present on the training scene?
2. Is evaluation focusing on results rather than on the effort expended in conducting training?
continued...
EVALUATION AUDIT
4. Is collection of data and interpretation of results done by personnel qualified for the job?
8.Is evaluation an aid to future planning, is it directive and constructive and not conclusive?
9. Do trainees participate in the evaluation of their own progress? Are evaluation procedures
reviewed and revised periodically?
10. Are tests/ examinations used, derived from training objectives and are consistent with the
coverage of inputs?
11. Are other methods like observation, ratings, opinion surveys, interviews used to supplement
tests?
12. Are scoring, grading and reporting practices standardized, economical, practical,
acceptable(SEPA)?
continued...
EVALUATION AUDIT
14. Is the evaluation exercise worth the time, money and effort?
ii)REPORT WRITING
COVER PAGE
Name of Organization
Title of Training
Name of Evaluator
Report Date
CONTENTS
Executive Summary
Introduction
Participant’s Reaction
Learning
Job Impact
Business Impact
Additional Information
Acknowledgements
References
1)Executive Summary
Purpose
Objectives
Methodology
Main findings
Key recommendations
2)INTRODUCTION
Background
Description of training
Training Objectives
Number starting/completing
Job roles/departments
Age
Gender
Background
3)PARTICIPANT’S REACTION
Summarise/insert data from your Participant’s Reaction evaluation(s), e.g. statistics, graphs and/or
lists of comments.
4)LEARNING
Summarise/insert data from your Learning evaluation(s), e.g. statistics, graphs and/or lists of
comments.
5)JOB IMPACT
Summarise/insert data from your Job Impact evaluation(s), e.g. statistics, graphs and/or lists of
comments.
6)BUSINESS IMPACT
Summarise/insert data from your from your Organisational Results evaluation(s), e.g.:
Additional information/comments
Eg. Summarise key factors in determining outcomes such as barriers and promoters of performance
improvement. Identify other factors which may have influenced performance outcomes.
Payback Period
Additional information/comments
Eg. Summarise any limitations on data and identify other factors which may have influenced ROTI
outcomes.
8)ADDITIONAL INFORMATION
Add any further information that you feel may be relevant to the impact evaluation.
Also consider the impacts in relation to the costs and benefits of the investment.
10)Acknowledgement
11) References