Lesson Seven Item Analysis. Contents Item Analysis Item Analysis Item difficulty (item facility) Item difficulty (item facility) Item difficulty Item.

Slides:



Advertisements
Similar presentations
Item Analysis.
Advertisements

Test Development.
FACULTY DEVELOPMENT PROFESSIONAL SERIES OFFICE OF MEDICAL EDUCATION TULANE UNIVERSITY SCHOOL OF MEDICINE Using Statistics to Evaluate Multiple Choice.
Item Analysis: Improving Multiple Choice Tests Crystal Ramsay September 27, 2011 Schreyer Institute for Teaching.
Lesson Six Reliability.
Using Test Item Analysis to Improve Students’ Assessment
Item Analysis: A Crash Course Lou Ann Cooper, PhD Master Educator Fellowship Program January 10, 2008.
Dr. Majed Wadi MBChB, MSc Med Edu
Some Practical Steps to Test Construction
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
Item Analysis What makes a question good??? Answer options?
Yun-Pi Yuan1 Lesson Ten Multiple-Choice Format. Yun-Pi Yuan2 Contents Advantages and disadvantages Terminology Guidelines for making items The stem The.
Item Analysis Prof. Trevor Gibbs. Item Analysis After you have set your assessment: How can you be sure that the test items are appropriate?—Not too easy.
Lesson Nine Test Evaluation: The JCEE Samples Contents ► How to Decide P and D How to Decide P and D How to Decide P and D ► Seeing-Eye Dogs Seeing-Eye.
Lesson Nine Item Analysis.
Lesson Three Kinds of Test and Testing. Yun-Pi Yuan 2 Contents Kinds of Tests: Based on Purposes  Classroom use Classroom use  External examination.
Multiple Choice Test Item Analysis Facilitator: Sophia Scott.
ANALYZING AND USING TEST ITEM DATA
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
Dunbar Vocational Career Academy High School Quarterly Interim Assessments & Data Driven Instruction SLCs & Departments.
Designing and evaluating good multiple choice items Jack B. Monpas-Huber, Ph.D. Director of Assessment & Student Information.
Chapter 8 Measuring Cognitive Knowledge. Cognitive Domain Intellectual abilities ranging from rote memory tasks to the synthesis and evaluation of complex.
Part #3 © 2014 Rollant Concepts, Inc.2 Assembling a Test #
Induction to assessing student learning Mr. Howard Sou Session 2 August 2014 Federation for Self-financing Tertiary Education 1.
Field Test Analysis Report: SAS Macro and Item/Distractor/DIF Analyses
Chapter 7 Item Analysis In constructing a new test (or shortening or lengthening an existing one), the final set of items is usually identified through.
Grading and Reporting Chapter 15
Techniques to improve test items and instruction
EDU 385 Education Assessment in the Classroom
Group 2: 1. Miss. Duong Sochivy 2. Miss. Im Samphy 3. Miss. Lay Sreyleap 4. Miss. Seng Puthy 1 ROYAL UNIVERSITY OF PHNOM PENH INSTITUTE OF FOREIGN LANGUAGES.
NRTs and CRTs Group members: Camila, Ariel, Annie, William.
Lesson Three Kinds of Test and Testing. Contents Kinds of Tests: Based on Purposes  Classroom use Classroom use  External examination Kinds of Testing:
Lab 5: Item Analyses. Quick Notes Load the files for Lab 5 from course website –
Item specifications and analysis
Basic Measurement and Statistics in Testing. Outline Central Tendency and Dispersion Standardized Scores Error and Standard Error of Measurement (Sm)
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
RELIABILITY AND VALIDITY OF ASSESSMENT
Writing Multiple Choice Questions. Types Norm-referenced –Students are ranked according to the ability being measured by the test with the average passing.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
A Study on Junior High Students ’ English Learning Achievement in Taiwan CamilaShelly January, 15, 2010.
Introduction to Item Analysis Objectives: To begin to understand how to identify items that should be improved or eliminated.
Review: Alternative Assessments Alternative/Authentic assessment Real-life setting Performance based Techniques: Observation Individual or Group Projects.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Tests and Measurements
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
Educational Research Chapter 8. Tools of Research Scales and instruments – measure complex characteristics such as intelligence and achievement Scales.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 6 Construction of Knowledge Tests.
Dept. of Community Medicine, PDU Government Medical College,
Objective Examination: Multiple Choice Questions Dr. Madhulika Mistry.
Norm Referenced Your score can be compared with others 75 th Percentile Normed.
Exam Analysis Camp Teach & Learn May 2015 Stacy Lutter, D. Ed., RN Nursing Graduate Students: Mary Jane Iosue, RN Courtney Nissley, RN Jennifer Wierworka,
COMMON TEST TECHNIQUES FROM TESTING FOR LANGUAGE TEACHERs.
Using Data to Drive Decision Making:
ARDHIAN SUSENO CHOIRUL RISA PRADANA P.
Data Analysis and Standard Setting
Classroom Analytics.
Greg Miller Iowa State University
Test Development Test conceptualization Test construction Test tryout
Classroom Assessment Ways to improve tests.
Item writing.
TOPIC 4 STAGES OF TEST CONSTRUCTION
Dept. of Community Medicine, PDU Government Medical College,
Using statistics to evaluate your test Gerard Seinhorst
Summative Assessment Grade 6 April 2018 Develop Revise Pilot Analyze
Lies, Damned Lies & Statistical Analysis for Language Testing
Analyzing test data using Excel Gerard Seinhorst
Distractor Efficiency
Timed Multiple Choice Questions
Tests are given for 4 primary reasons.
Timed Multiple Choice Questions
Presentation transcript:

Lesson Seven Item Analysis

Contents Item Analysis Item Analysis Item difficulty (item facility) Item difficulty (item facility) Item difficulty Item difficulty Item discrimination Item discrimination Item discrimination Item discrimination Distrctor (efficiency) analysis Distrctor (efficiency) analysis Distrctor (efficiency) analysis Distrctor (efficiency) analysis

Item Difficulty (1) Also called item facility (Bailey ) Also called item facility (Bailey ) Difficulty index: P or IF Difficulty index: P or IF In terms of how many people got this item correct (i.e., the percentage of Ss who correctly answer a given item). In terms of how many people got this item correct (i.e., the percentage of Ss who correctly answer a given item). P= The number of correct responses divided by the number of total test takers. P= The number of correct responses divided by the number of total test takers. P= N correct / N total = (%U+ %L) / 2 P= N correct / N total = (%U+ %L) / 2

Item Difficulty (2) 0 ≤ P ≤ 1 (0= difficult item, 1= easy item) 0 ≤ P ≤ 1 (0= difficult item, 1= easy item) Achievement Test (CRT): P=0 or P= 1  possible Achievement Test (CRT): P=0 or P= 1  possible NRT: P=0 or P=1  means the designer has to go back and check the question NRT: P=0 or P=1  means the designer has to go back and check the question How high should the P be? How high should the P be? Depending on purpose of giving the test Depending on purpose of giving the test If for mastery of the essential  P>=.8 If for mastery of the essential  P>=.8 If for spreading people out  P should be lower If for spreading people out  P should be lower

Item Discrimination (1) Discrmination index: D or ID Discrmination index: D or ID Indicate the degree to which the task differentiates correctly among test takers; the difference of the percentage passing between upper group and lower group Indicate the degree to which the task differentiates correctly among test takers; the difference of the percentage passing between upper group and lower group D= % of upper group - % of lower group OR D = D= % of upper group - % of lower group OR D = Correct Answer in Upper Group – Correct Answer in Lower Group No. of people in each group (one group)

Item Discrimination (2) How to interpret the discrimination index? -1<= D<= 1 How to interpret the discrimination index? -1<= D<= 1 When D= 0 (i.e., %U = %L) When D= 0 (i.e., %U = %L) ok/great  if testing for mastery ok/great  if testing for mastery terrible  if testing for spreading people out terrible  if testing for spreading people out D>=.25 or.35 (Bailey 135) D>=.25 or.35 (Bailey 135) D =.7 (realistic D): probably the best D that we could get D =.7 (realistic D): probably the best D that we could get

Item Discrimination (3) Definition of upper/lower group: Definition of upper/lower group: Sharper differentiation when using more extreme ends (e.g., 10% top & bottom) Sharper differentiation when using more extreme ends (e.g., 10% top & bottom) But, reduce reliability of the results (because of small # of cases utilized) But, reduce reliability of the results (because of small # of cases utilized) Q: What’s the optimum percentage (used to define upper & lower group) so that the characteristics of differentiation & reliability can be maximized? Q: What’s the optimum percentage (used to define upper & lower group) so that the characteristics of differentiation & reliability can be maximized?

Item Discrimination (4) The optimum percentage = 27% (Anastasi 1988) The optimum percentage = 27% (Anastasi 1988) Any convenient number between 25% ~ 33% will do. (Yuan 1998) Any convenient number between 25% ~ 33% will do. (Yuan 1998) The JCEE uses 33% The JCEE uses 33% Now, do exercises on Bailey 146. Now, do exercises on Bailey 146.

Distractor Analysis (1) Also called distractor efficiency analysis Also called distractor efficiency analysis To investigate the degree to which the distractors are functioning efficiently (i.e., attracting Ss who don’t know the correct answer). To investigate the degree to which the distractors are functioning efficiently (i.e., attracting Ss who don’t know the correct answer). Specifically related to the multiple- choice format. Specifically related to the multiple- choice format.

Distractor Analysis (2) Choices A B C D* Choices A B C D* U% U% L% L% P = ? D = ? P = ? D = ? P = (.8 +.5)/2 = 1.3/2 =.65 P = (.8 +.5)/2 = 1.3/2 =.65 D = =.3 D = =.3 Now, look at each of the distractors. Now, look at each of the distractors. A  negatively discriminative A  negatively discriminative C  no appeal at all C  no appeal at all

Distractor Analysis (3) Choices A B C* D Choices A B C* D high P=.6 high P=.6 low D= 0 low D= 0 Examine the original item carefully (in terms of its format and content) Examine the original item carefully (in terms of its format and content) Which choices should be revised? Which choices should be revised?

Distractor Analysis (4) Choices A* B C D Choices A* B C D high P=.5 high P=.5 low D= low D= What does the result of distractor analysis tell us? What does the result of distractor analysis tell us? The item might be miskeyed? Should B be the correct answer? The item might be miskeyed? Should B be the correct answer?