Item Analysis: Improving Multiple Choice Tests Crystal Ramsay September 27, 2011 Schreyer Institute for Teaching.

Slides:



Advertisements
Similar presentations
EU Presidency Conference Effective policies for the development of competencies of youth in Europe Warsaw, November 2011 Improving basic skills in.
Advertisements

Writing constructed response items
1 The Academic Profession and the Managerial University: An International Comparative Study from Japan Akira Arimoto Research Institute for Higher Education.
The Course experience questionnaire (P. Ramsden) Designed as a performance indicator 24 statements relating to 5 aspects 1 overall satisfaction statement.
APS Teacher Evaluation
The Framework for Teaching Charlotte Danielson
Chapter 2 Instructional Systems Presented By: Presented By: Shelly Romack Shelly Romack Instructional Technology and Media for Learning Smaldino, Russell,
The SCPS Professional Growth System
Copyright © 2007 Ramez Elmasri and Shamkant B. Navathe Slide
FASS Disability Initiative Seminar Two: Curriculum and Course Design Dr Leanne Dowse (SSIS) and Dr Brooke Dining.
Item Analysis.
How to Make a Test & Judge its Quality. Aim of the Talk Acquaint teachers with the characteristics of a good and objective test See Item Analysis techniques.
FACULTY DEVELOPMENT PROFESSIONAL SERIES OFFICE OF MEDICAL EDUCATION TULANE UNIVERSITY SCHOOL OF MEDICINE Using Statistics to Evaluate Multiple Choice.
(This presentation may be used for instructional purposes)
ASSESSMENT 101 Preparing Future Faculty (PFF) Workshop Spring 2011 Facilitators: Dr. Kara Penfield, Director of Academic Assessment; Isis Artze-Vega, PFF.
A Strategy to Determine Source of Noncompliance Issues and Improve Services Trainer: Marilyn Johnson, PhD March 17, 2011 Improvement of Student Performance.
1 Developing Tests for Departmental Assessment Deborah Moore, Assessment Specialist Institutional Research, Planning, & Effectiveness University of Kentucky.
Why Take EXPLORE? EXPLORE shows your strengths and weaknesses in English, mathematics, reading, and science. EXPLORE helps you search for careers and.
Copyright © 2014 by Educational Testing Service. ETS, the ETS logo, LISTENING. LEARNING. LEADING. and GRE are registered trademarks of Educational Testing.
1 Measure Up! Benchmark Assessment Quality Assurance Process RCAN September 10, 2010.
Summative Assessment Kansas State Department of Education ASSESSMENT LITERACY PROJECT1.
Using Test Item Analysis to Improve Students’ Assessment
Using Multiple Choice Tests for Assessment Purposes: Designing Multiple Choice Tests to Reflect and Foster Learning Outcomes Terri Flateby, Ph.D.
Item Analysis: A Crash Course Lou Ann Cooper, PhD Master Educator Fellowship Program January 10, 2008.
Some Practical Steps to Test Construction
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
Item Analysis What makes a question good??? Answer options?
Lesson Seven Item Analysis. Contents Item Analysis Item Analysis Item difficulty (item facility) Item difficulty (item facility) Item difficulty Item.
Item Analysis Prof. Trevor Gibbs. Item Analysis After you have set your assessment: How can you be sure that the test items are appropriate?—Not too easy.
Multiple Choice Test Item Analysis Facilitator: Sophia Scott.
ANALYZING AND USING TEST ITEM DATA
Stages of testing + Common test techniques
Chap. 3 Designing Classroom Language Tests
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Multiple Choice vs. Performance Based Tests in High School Physics Classes Katie Wojtas.
Writing Calibration English I Tuesday, September 2 nd, 2014.
Part #3 © 2014 Rollant Concepts, Inc.2 Assembling a Test #
1 Testing and Grading Ensuring that we meet the objectives.
TEST DESIGN Presented by: Danielle Harrison. INTRODUCTION  What is a test? “Any activity that indicates how well learners meet learning objectives is.
Chapter 7 Item Analysis In constructing a new test (or shortening or lengthening an existing one), the final set of items is usually identified through.
Techniques to improve test items and instruction
Session 2 Traditional Assessments Session 2 Traditional Assessments.
Assignment Week 4 Chandra Wilson EDU697: Capstone: A Project Approach Dr. Phillip Orlando May 20, 2013.
Group 2: 1. Miss. Duong Sochivy 2. Miss. Im Samphy 3. Miss. Lay Sreyleap 4. Miss. Seng Puthy 1 ROYAL UNIVERSITY OF PHNOM PENH INSTITUTE OF FOREIGN LANGUAGES.
NRTs and CRTs Group members: Camila, Ariel, Annie, William.
Lab 5: Item Analyses. Quick Notes Load the files for Lab 5 from course website –
Grading and Analysis Report For Clinical Portfolio 1.
1 Item Analysis - Outline 1. Types of test items A. Selected response items B. Constructed response items 2. Parts of test items 3. Guidelines for writing.
Common Formative Assessments for Science Monica Burgio Daigler, Erie 1 BOCES.
Educator’s view of the assessment tool. Contents Getting started Getting around – creating assessments – assigning assessments – marking assessments Interpreting.
Assessment and Testing
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Building Exams Dennis Duncan University of Georgia.
Introduction to Item Analysis Objectives: To begin to understand how to identify items that should be improved or eliminated.
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
Psychometrics: Exam Analysis David Hope
Assessment and the Institutional Environment Context Institutiona l Mission vision and values Intended learning and Educational Experiences Impact Educational.
Dept. of Community Medicine, PDU Government Medical College,
Norm Referenced Your score can be compared with others 75 th Percentile Normed.
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
Exam Analysis Camp Teach & Learn May 2015 Stacy Lutter, D. Ed., RN Nursing Graduate Students: Mary Jane Iosue, RN Courtney Nissley, RN Jennifer Wierworka,
ARDHIAN SUSENO CHOIRUL RISA PRADANA P.
Classroom Analytics.
Greg Miller Iowa State University
Dept. of Community Medicine, PDU Government Medical College,
Using statistics to evaluate your test Gerard Seinhorst
Analyzing test data using Excel Gerard Seinhorst
Learning Assessment Learning Teaching Dr. Md. Mozahar Ali
Tests are given for 4 primary reasons.
Presentation transcript:

Item Analysis: Improving Multiple Choice Tests Crystal Ramsay September 27, 2011 Schreyer Institute for Teaching Excellence

This workshop is designed to help you do three things: To interpret statistical indices provided by the universitys Scanning Operations To differentiate between well-performing items and poor-performing items To make decisions about poor performing items

We give tests for 4 primary reasons. To find out if students learned what we intended To separate those who learned from those who didnt To increase learning and motivation To gather information for adapting or improving instruction

The rounded filling of an internal angle between two surfaces of a plastic molding is known as the A.rib. B.fillet. C.chamfer. D.Gusset plate. Stem Distracters Key Options Multiple choice items are comprised of 4 basic components.

An item analysis focuses on 4 major pieces of information provided in the test score report. Test Score Reliability Item Difficulty Item Discrimination Distracter information

Test score reliability is an index of the likelihood that scores would remain consistent over time if the same test was administered repeatedly to the same learners. Reliability coefficients range from.00 to Now look at the test score reliability from your exam. Ideal score reliabilities are >.80. Higher reliabilities = less measurement error.

Item Difficulty is the percentage of students who answered an item correctly. RESPONSE TABLE - FORM AITEM NO. OMIT A B C D E KEY- % EFFECT % % % C A C Represented in the Response Table as KEY-% Ranges from 0% to 100%

Easier items have higher item difficulty values. More difficult items have lower item difficulty values. ITEM NO. RESPONSE TABLE –FORM A ITEM EFFECT OMITABCDEKEY -% %%% C A E ITEM NO. RESPONSE TABLE –FORM A ITEM EFFECT OMITABCDEKEY -% %%% D D D

Number of alternatives for each item What is an ideal item difficulty statistic depends on 2 factors. Your reason for asking the question

Sometimes we include very easy or very difficult items on purpose. Did I deliberately pose difficult items to challenge my students thinking? Did I deliberately pose easy items to test basic information or to boost students confidence?

Now look at the item difficulties from your exam. Which items were easier for your students? Which items were more difficult?

Item Discrimination is the degree to which students with high overall exam scores also got a particular item correct. RESPONSE TABLE - FORM AITEM NO. OMIT A B C D E KEY- % EFFECT % % % C A C Represented as Item Effect because it tells how well an item performed Ranges from to 1.00 and should be >.2

A well- performing item A poor- performing item ITEM NO. RESPONSE TABLE –FORM A ITEM EFFECT OMITABCDEKEY -% %%% E ITEM NO. RESPONSE TABLE –FORM A ITEM EFFECT OMITABCDEKEY -% %%% D 0.46

Item Difficulty Test heterogeneity Item characteristics What is an ideal item discrimination statistic depends on 3 factors.

Very easy or very difficult items will have poor ability to discriminate among students. Very easy or very difficult items may still be necessary to sample content taught. Yet… Item difficulty

A test that assesses many different topics will have a lower correlation with any one content-focused item. A heterogeneous item pool may still be necessary to sample content taught. Yet… Test heterogeneity

A poorly written item will have little ability to discriminate among students. There is no substitute for a well-written item or for testing what you teach! and… Item quality

Now look at the item effects from your exam. Which items on your exam performed well? Did any items perform poorly?

Distracter information can be analyzed to determine which distracters were effective and which ones were not. RESPONSE TABLE - FORM AITEM NO. OMIT A B C D E KEY- % EFFECT % % % C A C Now look at the distracter information for items from your exam. What can you conclude about them?

Whether to retain, revise, or eliminate items depends on item difficulty, item discrimination, distracter information, and your instruction. Distracters Ultimately, its a judgment call that you have to make. Instruction

What if I have a relatively short test or I give a test in a small class? I might not use the testing service for scoring. Is there a way I can understand how my items worked? Yes.

Item 1AB*CD Top 1/310 Bottom 1/31432 Item 2A*BCD Top 1/382 Bottom 1/373 Item 3ABC*D Top 1/3514 Bottom 1/3244 Item 4A*BCD Top 1/310 Bottom 1/391 From: Suskie, L. (2009). Assessing student learning: A common sense guide (2 nd ed.). San Francisco: Jossey-Bass. 1. Which item is the easiest? 2. Which item shows negative (very bad) discrimination? 3. Which item discriminates best between high and low scores? 4. In Item 2, which distracter is most effective? 5. In Item 3, which distracter must be changed?

Multiple course sections Student feedback Other item types Even after you consider reliability, difficulty, discrimination, and distracters, there are still a few other things to think about…

Resources For an excellent resource on item analysis: eport/itemanalysis.php eport/itemanalysis.php For a more extensive list of item-writing tips: Choice%20Item%20Writing%20Guidelines%20- %20Haladyna%20and%20Downing.pdf Choice%20Item%20Writing%20Guidelines%20- %20Haladyna%20and%20Downing.pdf c_tips.pdf c_tips.pdf For a discussion about writing higher-level multiple choice items: dford.pdf dford.pdf