Training Evaluation.

Slides:



Advertisements
Similar presentations
Training Companies are in business to make money, and every business function is under pressure to show how it contributes to business success or face.
Advertisements

Training. Training & Development Definition “The systematic acquisition of attitudes, concepts, knowledge, roles, or skills, that result in improved performance.
Evaluation Procedures
Reasons for Evaluating Training Companies are investing millions of dollars in training programs to help gain a competitive advantage. To justify the costs.
Defining Characteristics
GROUP-LEVEL DESIGNS Chapter 9.
The Basics of Experimentation I: Variables and Control
Experimental Research Designs
Research Design: The Experimental Model and Its Variations
Introduction to Research Design Threats to Internal Validity Two or More Groups Social Threats.
Research Design and Validity Threats
Unit 10: Evaluating Training and Return on Investment 2009.
Training & Development Definition –“The systematic acquisition of attitudes, concepts, knowledge, roles, or skills, that result in improved performance.
Evaluation of Training
6 Chapter Training Evaluation.
9 Quantitative Research Designs.
McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc. All rights reserved Chapter Training Evaluation.
Evaluating Training Programs. How can training programs be evaluated? Measures used in evaluating training programs Measures used in evaluating training.
Chapter 6 Training Evaluation
Human Resources Training and Individual Development February 11: Training Evaluation.
Chapter 9 Experimental Research Gay, Mills, and Airasian
Experimental Research
1.
Chapter 8 Experimental Research
Chapter 5 Training Discuss how training can contribute to companies’ business strategy. Explain the role of the manager in identifying training needs and.
Experimental Design The Gold Standard?.
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Primary Data Collection: Experimentation CHAPTER eight.
Group Discussion Explain the difference between assignment bias and selection bias. Which one is a threat to internal validity and which is a threat to.
Day 6: Non-Experimental & Experimental Design
Human Resource Management Gaining a Competitive Advantage
Chapter 11 Experimental Designs
Chapter 6 Training Evaluation. Chapter 6 Training Evaluation Concepts Training Evaluation: The process of collecting data regarding outcomes needed to.
5 Chapter Training Evaluation.
Learning Objectives Copyright © 2004 John Wiley & Sons,Inc Primary Data Collection: Experimentation CHAPTER Seven.
Training Evaluation. Evaluation of Training Since huge sums of money are spent on training and development, how far the programme has been useful must.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Primary Data Collection: Experimentation CHAPTER eight.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Evaluating HRD Programs
Dr. Shulagna Sarkar Assistant Professor , Jt. Coordinator – PGDM – HRM
Chapter 6 Training Evaluation
Chapter 10 Experimental Research Gay, Mills, and Airasian 10th Edition
Chapter Six: The Basics of Experimentation I: Variables and Control.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Research Design. Time of Data Collection Longitudinal Longitudinal –Panel study –Trend study –Cohort study Cross-sectional Cross-sectional.
Chapter Eight: Quantitative Methods
11-1 Chapter 11 Experiments and Test Markets Learning Objectives Understand... uses for experimentation advantages and disadvantages of the experimental.
BBA 229 Training and Development
Evaluation of Training Rationale for Evaluation Types of Evaluation Data Validity Issue Evaluation Design.
CHAPTER 8 EXPERIMENTS.
Advances in Human Resource Development and Management Course Code: MGT 712 Lecture 29.
Behrens Automotive Components Company Pre-International Assignment Training Proposal.
Training Evaluation Chapter 6
Training Evaluation Chapter 6 6 th Edition Raymond A. Noe Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
6 - 1 Training Evaluation Introduction (1 of 2) Training effectivenessTraining effectiveness refers to the benefits that the company and the trainees.
School of Public Administration & Policy Dr. Kaifeng Yang 研究设计 : 实验研究的基本问题.
Research designs Research designs Quantitative Research Designs.
Educational Research Experimental Research Chapter 9 (8 th Edition) Chapter 13 (7 th Edition) Gay and Airasian.
EXPERIMENTAL RESEARCH
Chapter Six Training Evaluation.
Chapter Six Training Evaluation.
Chapter Eight: Quantitative Methods
SPIRIT OF HR.in TRAINING EVALUATION.
Purpose of Training Evaluation
Chapter 18: Experimental and Quasi-Experimental Research
Group Experimental Design
Chapter 11 EDPR 7521 Dr. Kakali Bhattacharya
Training Evaluation Chapter 6
6 Chapter Training Evaluation.
Reminder for next week CUELT Conference.
Presentation transcript:

Training Evaluation

Training evaluation provides the data needed to demonstrate that training does provide benefits to the company.

What are the differences among: Training effectiveness Training outcomes Training evaluation Evaluation design Training effectiveness refers to the benefits that the company and the trainees receive from training Training outcomes or criteria refer to measures that the trainer and the company use to evaluate training programs Training evaluation refers to the process of collecting the outcomes needed to determine if training is effective Evaluation design refers to from whom, what, when, and how information needed for determining the effectiveness of the training program will be collected

Types of Evaluation Formative Summative Formative evaluation – evaluation conducted to improve the training process Helps to ensure that: the training program is well organized and runs smoothly trainees learn and are satisfied with the program Provides information about how to make the program better Summative evaluation – evaluation conducted to determine the extent to which trainees have changed as a result of participating in the training program May also measure the return on investment (ROI) that the company receives from the training program

Why Evaluate Training Programs? Objectives Satisfaction Benefits Comparison Why Evaluate Training Programs Objectives to determine whether the program is meeting objectives, is enhancing learning, and is resulting in transfer of training to the job Satisfaction to determine whether trainees believe that the content and administration of the program were satisfactory Benefits to determine the financial benefits and costs of the program Comparison to compare the costs and benefits of different training programs to choose the best program

Why Should A Training Program Be Evaluated? (1 of 2) To identify the program’s strengths and weaknesses To assess whether content, organization, and administration of the program contribute to learning and the use of training content on the job To identify which trainees benefited most or least from the program

Why Should A Training Program Be Evaluated? (2 of 2) To gather data to assist in marketing training programs To determine the financial benefits and costs of the programs To compare the costs and benefits of training versus non-training investments To compare the costs and benefits of different training programs to choose the best program

Objectives = Foundation Terminal behavior Conditions under which terminal behavior is expected The standard below which performance is unacceptable --> criteria by which the trainee is judged

Training Program Objectives and Their Implications for Evaluation: Learning Transfer Outcomes Did company benefit through sales, quality, productivity, reduced accidents, and complaints? Performance on work equipment Results: Performance on a work sample Skill-Based: Trainees’ motivation or job attitudes Affective: Pencil-and-paper tests Cognitive: Ratings by peers or managers based on observation of behavior Did trainees like the program? Did the environment help learning? Was material meaningful? Reactions:

The Evaluation Process Conduct a Needs Analysis Develop Measurable Learning Outcomes and Analyze Transfer of Training Develop Outcome Measures Choose an Evaluation Strategy Plan and Execute the Evaluation

Training Outcomes: Kirkpatrick’s Four-Level Framework of Evaluation Criteria Business results achieved by trainees Results 4 Improvement of behavior on the job; aka skills Behavior 3 Acquisition of knowledge, skills, attitudes, behavior; aka cognitive Learning 2 Trainee satisfaction; aka affective Reactions 1 Focus Criteria Level

Training Outcomes Cognitive (Knowledge) Skills (Behaviors) Affect (Attitudes & Motivation) Reactions Results reaction outcomes relate to what trainees thought of the program, including the facilities, trainers, & content >survey of participants asking what they thought about the program learning outcomes are used to determine the extent to which trainees are familiar with principles, facts, techniques, skills, or processes presented in the training program >administration of an exam over training materials behavior outcomes are used to determine the extent to which the trainees’ behavior changes on the job as a result of participation in the program >recording behavior on the job after training and comparing it to behavior on the job before training results outcomes are used to determine the payoff the training program had for the company >measuring productivity after training to see if it increased

Which Outcome is Represented by the Evaluation Form for Your Training Module?

Outcomes Used in Evaluating Training Programs: (1 of 4) Affective Outcomes Results Return on Investment Cognitive Skill-Based The following suggestions specify the types of changes in jobs that are most likely to lead to improvements in each of the five core dimensions. (1) Combine tasks - managers should put existing fractionalized tasks back together to form a new, larger module of work. This increases skill variety and task identify. (2) Create natural work units - managers should design tasks that form an identifiable and meaningful whole. This increases employee “ownership” of the work and encourages employees to view their work as meaningful and important rather than as irrelevant and boring. (3) Establish client relationships - the client is the user of the product or service that the employee works on. Whenever possible, managers should establish direct relationships between workers and their clients. This increases skill variety, autonomy, and feedback for the employee. (4) Expand jobs vertically - vertical expansion means giving employees responsibilities and controls that were formerly reserved for management. It partially closes the gap between the “doing” and “controlling” aspects of the job, and it increases employee autonomy. (5) Open feedback channels - by increasing feedback, employees not only learn how well they are performing their jobs but also whether their performances are improving, deteriorating, or remaining at a constant level. Ideally, employees should receive performance feedback directly as they do their jobs rather than from management on an occasional basis. 9

Outcomes Used in Evaluating Training Programs: (2 of 4) Cognitive Outcomes Determine the degree to which trainees are familiar with the principles, facts, techniques, procedures, or processes emphasized in the training program Measure what knowledge trainees learned in the program Skill-Based Outcomes Assess the level of technical or motor skills Include acquisition or learning of skills and use of skills on the job

Outcomes Used in Evaluating Training Programs: (3 of 4) Affective Outcomes Include attitudes and motivation Trainees’ perceptions of the program including the facilities, trainers, and content Results Determine the training program’s payoff for the company

Outcomes Used in Evaluating Training Programs: (4 of 4) Return on Investment (ROI) Comparing the training’s monetary benefits with the cost of the training direct costs indirect costs benefits

How do you know if your outcomes are good? Good training outcomes need to be: Relevant Reliable Discriminative Practical

Good Outcomes: Relevance Criteria relevance – the extent to which training programs are related to learned capabilities emphasized in the training program Criterion contamination – extent that training outcomes measure inappropriate capabilities or are affected by extraneous conditions Criterion deficiency – failure to measure training outcomes that were emphasized in the training objectives

Criterion deficiency, relevance, and contamination: Outcomes Identified by Needs Assessment and Included in Training Objectives Outcomes Related to Training Objectives Outcomes Measured in Evaluation Contamination Relevance Deficiency

Good Outcomes (continued) Reliability – degree to which outcomes can be measured consistently over time Discrimination – degree to which trainee’s performances on the outcome actually reflect true differences in performance Practicality – refers to the ease with which the outcomes measures can be collected

Training Evaluation Practices Percentage of Courses Using Outcome Outcomes

How do Fidelity & Motorola evaluate their training programs? How do their measures of success compare with those advocated by the text?

Evaluation Designs: Threats to Validity Threats to validity refer to a factor that will lead one to question either: The believability of the study results (internal validity), or The extent to which the evaluation results are generalizable to other groups of trainees and situations (external validity)

Threats to Validity Threats To External Validity Company Threats To Internal Validity Company Persons Outcome Measures Threats To External Validity Reaction to pretest Reaction to evaluation Interaction of selection and training Interaction of methods

Methods to Control for Threats to Validity Pre- and Posttests Use of Comparison Groups Random Assignment

Evaluation Procedures

Utility

[(Ns)*(T)*(r)*(SDy)*(Zs)]-[(N)*(C)] Ns = number of applicants selected T = tenure of selected group in years r = correlation between predictor and job performance (VALIDITY) SDy = standard deviation of job performance Zs = average standard predictor score of selected group N = number of applicants C = cost per applicant

[(Nc)*(T)*(r)*(SDy)*(Zs)]-[(N)*(C)] Nc = number of trainees who complete program T = duration of training benefit r = correlation between training criterion and job performance (VALIDITY) SDy = standard deviation of job performance Zs = average standard criterion score of trainees N = total number of trainees enrolled C = cost per trainee

Training Costs Direct Indirect Development Overhead Compensation for Trainees Direct instructor (daily rate x # days, fringe benefits) travel expenses, materials x # trainees, classroom space & av equipment, refreshments (X #days X # trainees) Indirect mgmt, clerical & admin salaries + fringes; postage, shipping, telephone, pre & post learning materials (x # trainees) Development fee for program purchase, instructor training, registration fee, travel & lodging, salary, benefits Overhead general org support, top mgmt time (10% of d, I & d) Compensation for Trainees trainees salaries & benefits (based on time away from the job)

For On the Job Training $81,000 50 = Ns = number of trainees who complete program 1 = T = duration of training benefit .50 = r = correlation between training criterion and job performance (VALIDITY) 4800 = SDy = standard deviation of job performance (assume 40% of base pay . . . $12,000 * .40) .80 = Zs = average standard criterion score of trainees 100 = N = total number of trainees enrolled 150 = C = cost per trainee [(Ns)*(T)*(r)*(SDy)*(Zs)]-[(N)*(C)] (50 * 1 * .50 * 4800 * .8) - (100 * 150)

Experimental Designs

Experimental Designs Choices Pretest/posttest Control Groups

Experimental Designs 1: 1 group, posttest only 2: 1 group, pretest/posttest 3: Pretest/posttest control group 4: Solomon four-group 5: Time-series 6: Nonequivalent control group

Experimental Designs Validity Internal External

Experimental Designs Threats to Internal Validity History Maturation Testing Instrumentation Regression toward the mean Differential selection Experimental mortality Interactions Diffusion/imitation of treatments Compensatory equalization of treatments Rivalry/desirability of treatments Demoralization

Experimental Designs Threats to External Validity Reactive effect of pretesting Interaction of selection & treatment Reactive effects of experimental settings Multiple-treatment interference

Issues in Training Validity Transfer validity Intra-organizational validity Inter-organizational validity