CLEAR 2011 Annual Educational Conference

Slides:



Advertisements
Similar presentations
Knowledge Dietary Managers Association 1 PART II - DMA Certification Exam Blueprint and Exam Development-
Advertisements

Principles of Standard Setting
Test Development.
Unraveling the Mysteries of Setting Standards and Scaled Scores Julie Miles PhD,
Copyright © 2012 Pearson Education, Inc. or its affiliate(s). All rights reserved
M AKING A PPROPRIATE P ASS- F AIL D ECISIONS D WIGHT H ARLEY, Ph.D. DIVISION OF STUDIES IN MEDICAL EDUCATION UNIVERSITY OF ALBERTA.
2007 Annual Conference “ Windows and Continuous CBT: a Brief Overview ” George T Gray, EdD ACT, Inc.
Advanced Topics in Standard Setting. Methodology Implementation Validity of standard setting.
Medical school attendedPassing grade Dr JohnNorthsouth COM (NSCOM)80% Dr SmithEastwest COM (EWCOM)50% Which of these doctors would you like to treat you?
Standard Setting for Professional Certification Brian D. Bontempo Mountain Measurement, Inc. (503) ext 129.
1 New England Common Assessment Program (NECAP) Setting Performance Standards.
“You Can’t Always Get What You Want: Accountability and Limits in Score Reports to Boards” Troy Elliott Association of Social Work Boards 2006 Annual ConferenceAlexandria,
You’ve Been Shown: Now It’s Your Turn to Ask! CLEAR 2004Kansas City, Missouri Responding to Your Questions About Testing.
The Bookmark Procedure Your Logo Here. Basics Arrange test items from easy to hard Examine each item in order Consider the likelihood of a correct response.
Setting Performance Standards Grades 5-7 NJ ASK NJDOE Riverside Publishing May 17, 2006.
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
SETTING & MAINTAINING EXAM STANDARDS Raja C. Bandaranayake.
Setting Alternate Achievement Standards Prepared by Sue Rigney U.S. Department of Education NCEO Teleconference March 21, 2005.
Uses of Language Tests.
Examing Rounding Rules in Angoff Type Standard Setting Methods Adam E. Wyse Mark D. Reckase.
Wastewater Treatment Plant Operator Exam Setting Performance Standards With The Modified Angoff Procedure.
Standardized Test Scores Common Representations for Parents and Students.
Presenters: Promoting Regulatory Excellence ASK THE EXPERTS Sandra Greenberg, PhD Professional Examination Service Joseph McClintock, PhD Measurement Incorporated.
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Modified for EPE/EDP 711 by Kelly Bradley on January 8, 2013.
Standard Setting Methods with High Stakes Assessments Barbara S. Plake Buros Center for Testing University of Nebraska.
Kaizen–What Can I Do To Improve My Program? F. Jay Breyer, Ph.D. Presented at the 2005 CLEAR Annual Conference September Phoenix,
1 Establishing A Passing Standard Paul D. Naylor, Ph.D. Psychometric Consultant.
Overview of Standard Setting Leslie Wilson Assistant State Superintendent Accountability and Assessment August 26, 2008.
1 New England Common Assessment Program (NECAP) Setting Performance Standards.
CLEAR 2008 Annual Conference Anchorage, Alaska “Problems and Priorities in Pretesting Pondered” Beth Noeller Grady Barnhill Carol O’Byrne.
 Closing the loop: Providing test developers with performance level descriptors so standard setters can do their job Amanda A. Wolkowitz Alpine Testing.
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
Setting Cut Scores on Alaska Measures of Progress Presentation to Alaska Superintendents Marianne Perie, AAI July 27, 2015.
Classroom Diagnostic Tools. Pre-Formative Assessment of Current CDT Knowledge.
Standard Setting Results for the Oklahoma Alternate Assessment Program Dr. Michael Clark Research Scientist Psychometric & Research Services Pearson State.
Assessment and Performance Standards How Good is Good Enough? March 4-6, 2008.
Assessment and Testing
Using the Many-Faceted Rasch Model to Evaluate Standard Setting Judgments: An IllustrationWith the Advanced Placement Environmental Science Exam Pamela.
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Heriot Watt University 12th February 2003.
Unraveling the Mysteries of Setting Standards and Scaled Scores Julie Miles PhD,
NAEP Achievement Levels Michael Ward, Chair of COSDAM Susan Loomis, Assistant Director NAGB Christina Peterson, Project Director ACT.
Approach to Written Questions in MRCPCH Exam ELBABA 2012.
 Good for:  Knowledge level content  Evaluating student understanding of popular misconceptions  Concepts with two logical responses.
Every Student Succeeds Act (ESSA) Accountability
Designing Scoring Rubrics
Jean-Guy Blais Université de Montréal
What is a CAT? What is a CAT?.
Assessment of Learning 1
Assessments for Monitoring and Improving the Quality of Education
ARDHIAN SUSENO CHOIRUL RISA PRADANA P.
Wisconsin Department of Public Instruction
EDU 385 Session 8 Writing Selection items
Chapter 14 Assembling, Administering, and Appraising classroom tests and assessments.
Item Analysis: Classical and Beyond
Release of PARCC Student Results
The All-important Placement Cut Scores
Next-Generation MCAS: Update and review of standard setting
Multi Rater Feedback Surveys FAQs for Participants
Multi Rater Feedback Surveys FAQs for Participants
Preparing Tips For CompTIA SY0-501 Final Exam | CompTIA SY0-501 Dumps PDF Dumps4download
MANA 4328 Dr. Jeanne Michalski
Week 10 Slides.
The Processes and Requirements for Developing and Maintaining a Certification Program for Safety Professionals Presenter: David B. West, CSP, ASP, PE,
Standard Setting for NGSS
Rubrics for academic assessment
Standard Setting Zagreb, July 2009.
Item Analysis: Classical and Beyond
REVIEW I Reliability scraps Index of Reliability
Deanna L. Morgan The College Board
Item Analysis: Classical and Beyond
Presentation transcript:

CLEAR 2011 Annual Educational Conference September 8-10 Who Gets to Pass? Setting Defensible Passing Scores to Protect the Public Thomas Langenfeld: ACT, Inc Grady Barnhill: Commission on Dietetic Registration Jeff Kenney: National Council for Interior Design Qualification Pittsburgh, Pennsylvania

Goals of Today’s Program on Passing Scores: CLEAR 2011 Annual Educational Conference September 8-10 Goals of Today’s Program on Passing Scores: Through hands-on participation, we will learn the similarities and differences between three methods for determining a cut score. By participating, we will experience the cognitive requirements of each method. Have fun. Pittsburgh, Pennsylvania

Problems Inherent in all Pass Score Studies Application of Human Judgment – all methods entail a subjective process Panel Effect – different panels may produce different outcomes Method Effect – different methods may produce different outcomes What are we to do?

Common Requirements of Each Procedure Common Starting Point: Judges must fully understand the purpose and meaning of the credential. The test content must be carefully tied to requirements and meaning of the credential. Concept of qualification (competence): Judges must understand the concept of minimum qualification (competence)

Group Participation Hands-on participatory program Fictional Test Fictional Credential We will apply three passing score methods Modified Angoff Method Bookmark Method Hofstee Method

Fictional Test and Credential Purpose of Test - to certify state regulators in interpreting testing concepts to assist them in providing information to the public. Test: Interpreting Test Concepts Sponsoring Agency: NSPCP – The National Society for the Prevention of Cruelty to Psychometricians

Modified Angoff Procedure & Bookmark Procedure Concept of Minimum Competency (aka: the Just Qualified Candidate, the Minimally Qualified Candidate, the Barely Proficient Candidate) What knowledge would you expect? What skills would you expect? How does minimum competency relate to the task list? How good is good enough?

Working with Judges on Applying the concept of Minimum Competency CLEAR 2011 Annual Educational Conference September 8-10 Working with Judges on Applying the concept of Minimum Competency Focus on the Just Qualified Candidate Emphasize – this is the person who is just barely able to qualify. The Just Qualified Candidate – what do you expect him or her to know? To be able to do? Pittsburgh, Pennsylvania

Passing Standard (JTS 14.17) CLEAR 2011 Annual Educational Conference September 8-10 Passing Standard (JTS 14.17) The level of performance required for passing a credentialing test should depend on the knowledge and skills necessary for acceptable performance in the occupation or profession and should not be adjusted to regulate the number or proportion of persons passing the test. Just say no to Normative! Pittsburgh, Pennsylvania

Passing Point Considerations CLEAR 2011 Annual Educational Conference September 8-10 Passing Point Considerations Restriction of Trade vs. Public Protection Public Protection vs. Access to Care Adequate/inadequate vs. good/superior Restriction of trade Access to Care Public Protection Pittsburgh, Pennsylvania

Angoff Item Ratings How many out of 100 Just Qualified Examinees would you expect to answer each question correctly? Rate in increments of five. Comment: Even on extremely easy items, some examinees will misread, mismark, or make some other mistake in answering. Comment: Even on extremely difficult items, some examinees may guess the answer correctly.

Modified Angoff Process Judges go through items, one-by-one, and rate them Discussion Round 2: p-values provided; judges rate again Finalize ratings

Angoff Results Modified Angoff Table of Item Ratings 63.00 Items Diff 1 2 3 4 5 Avg. 0.70 60 45 70 66.43 0.47 50 65.36 0.63 75 55 66.07 0.56 67.86 0.42 65 40 6 0.77 80 73.00 Findings   63.33 60.83 65.00 68.33 57.50 63.00 4.11

Bookmark Procedure Book of Test Items: Ordered Item Booklet (OIB) In the OIB, items are arranged in order of difficulty from the least difficult to the most difficult. (Item Difficulty calibrated through an IRT model.) Because the Bookmark Procedure incorporates IRT, it possesses greater flexibility than other methods (i.e., it can be applied to a test using both multiple choice and constructed response items). Recommended passing score is on an IRT scale

Bookmark Directions for Judges Begin by reading the first item in the OIB Answer the question: Is it likely that a just qualified candidate will correctly answer the item? 50% or 67% Likely Judges answer “Yes” or “No” Proceed through the OIB answering “yes” or “no” for each item. When judges respond to an the item with “no,” stop and place the bookmark following the last “yes” item.

Bookmark Results Summary Data Table Raters Bookmark Placement IRT Theta Value 1 6 -0.84 2 7 -0.22 3 4 8 0.05 5 Results -0.41

Hofstee Method – Compromise Method Question #1 Based on my review of this set of test questions, I think that the percentage of correctly answered questions required to achieve certification should not go below __a____% correct, or above ___b____% correct. If I were to pick a single number as the requirement for passing, it would be ___c____% correct.

Hofstee Method Question No. 2 Based on my experience with this field of practice, I would not expect the overall failure rate for state regulators taking the Interpreting Test Concepts test to go below ____a___% fail, or be above ___b___% fail. If I were to select a single number as the expected failure rate for this group, it would be ___c___%.

Hofstee Procedure Calculating the Pass Score Calculate the average of each of the six responses. Two critical points: Failmin, Knowledgemax Failmax, Knowledgemin 3. Plot points and draw line across distribution.

Hofstee Example

Findings from the three procedures Findings for our Test Modified Angoff: Bookmark: Hofstee:

Assumptions of All Three Procedures Made Regarding the Judges Raters know the examinee population Raters know the test content Raters know the qualifications of the Just Qualified Examinees

Similarities and Differences Between the Methods Cognitive Requirements of Modified Angoff Cognitive Requirements of Bookmark Cognitive Requirements of Hofstee

Utilizing the Findings Setting the Final Pass Score Incorporating Results from Multiple Methods which may not provide the same passing standard Considering the consequences of False Positives and False Negatives (Consequential Validity) Ultimate responsibility for the performance standard resides with the organization granting the credential

CLEAR 2011 Annual Educational Conference September 8-10 Speaker Contact Information Thomas Langenfeld Jeff Kenney ACT, Inc. NCIDQ Iowa City, IA 52243 1602 L Street NW, Suite 200 Tel (319) 337-1125 Washington, DC 20036 Email: thomas.langenfeld@act.org Tel: (202)721-0220 Email: jeff@ncidq.org Grady Barnhill Commission on Dietetic Registration Chicago, IL 60606 Tel: (312) 899-4768 Email: gbarnhill@eatright.org Pittsburgh, Pennsylvania