Reflections on the Migration of Medical Simulators from Training Tools to Assessment Tools Rich Shavelson Stanford Education Assessment Lab Invited Talk.

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Data Collection Strategies An Overview. How Do We Assess ?
Instructional Technology vs. Educational Technology
Lecture 7: reliability & validity Aims & objectives –This lecture will explore a variety of techniques for ensuring that research is conducted with reliable.
MCR Michael C. Rodriguez Research Methodology Department of Educational Psychology.
Item Response Theory in a Multi-level Framework Saralyn Miller Meg Oliphint EDU 7309.
L2 program design Content, structure, evaluation.
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Psychological Testing Principle Types of Psychological Tests  Mental ability tests Intelligence – general Aptitude – specific  Personality scales Measure.
Part II Sigma Freud & Descriptive Statistics
Part II Sigma Freud & Descriptive Statistics
MEQ Analysis. Outline Validity Validity Reliability Reliability Difficulty Index Difficulty Index Power of Discrimination Power of Discrimination.
 Not Just a Collection of “Fill in the Blanks” Templates  Organized by Business Process  User Needs Assessment  Request for Proposal (RFP)  Vendor.
Miss. Mona AL-Kahtani.  Why do we test the oral ability? because we want to measure the development of the spoken language and the ability to interact.
Can you do it again? Reliability and Other Desired Characteristics Linn and Gronlund Chap.. 5.
Consistency/Reliability
Testing for Language Teachers
EVALUATION OF DR.MOHAMMED AL NAAMI, FRCSC, FACS, M Ed. Using O bjective S tructured C linical E xamination (OSCE)
How was the course organized? Faculty present: Lectures, small groups, and labs. Students prepare: Problem-based learning issues, journal article reports,
Classroom Assessment A Practical Guide for Educators by Craig A
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Good Customer Service Needs Good People Management.
EPortfolios: Getting started with Mahara. Ambition in Action ePortfolios:Getting started with Mahara o What is an ePortfolio o Examples.
Systems Theory and Integration Design Eliza Neumer.
NEXT GENERATION BALANCED ASSESSMENT SYSTEMS ALIGNED TO THE CCSS Stanley Rabinowitz, Ph.D. WestEd CORE Summer Design Institute June 19,
OJJDP Performance Measurement Training 1 Presented by: Dr. Kimberly Kempf-Leonard School of Social Sciences University of Texas at Dallas
Performance Samples Patricia L. Sitlington, Ph.D. University of Northern Iowa.
Assessing clinical judgment using the script concordance test: The importance of using specialty-specific experts to develop the scoring key Petrucci AM.
Presenter : Ching-ting Lin Instructor: Ming-puu Chen Developing a Usability Evaluation Method for E-learning Application: From Functional Usability to.
David W. Dillard AVCTC. Objectives Overview of the need for student assessments Define Student Assessments & parts of a rubric Samples of rubrics Develop.
Is the Script-Concordance Test a Valid Instrument for Assessment of Intra-operative Decision-making Skills? Brent Zabolotny 1, Robert Gagnon 2, Bernard.
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
1 Chapter 24 Scale Development and Statistical Analysis Methods for Scale Data.
General Information Iowa Writing Assessment The Riverside Publishing Company, 1994 $39.00: 25 test booklets, 25 response sheets 40 minutes to plan, write.
Validity and Reliability THESIS. Validity u Construct Validity u Content Validity u Criterion-related Validity u Face Validity.
Quality Assessment July 31, 2006 Informing Practice.
Identify a Health Problem Qualitative Quantitative Develop Program -theory -objectives -format -content Determine Evaluation -design -sampling -measures.
Selected Teaching-Learning Terms: Working Definitions...
International Seminar in Education Quality Control Standards in Medical Higher Educational Institutions.
THE MEASUREMENT OF USER INFORMATION SATISFACTION (BLAKE IVES ET.AL) Presented by: IRA GERALDINA
Assessment Module 5B ESUHSD June Outcomes Increase understanding of the Common Core State Standards (CCSS) in Mathematics by exploring assessment.
~ Test Construction and Validation ~
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Assessment Ice breaker. Ice breaker. My most favorite part of the course was …. My most favorite part of the course was …. Introduction Introduction How.
Welcome! A-Team Session 4 Fall Overview for Today 1 Minute Message review Outcomes Feedback Methodology Review Assessment Steps  Sampling  Validity/Reliability.
Writing A Review Sources Preliminary Primary Secondary.
Standards-Based Tests A measure of student achievement in which a student’s score is compared to a standard of performance.
METHODOLOGICAL STUDIES
MATHEMATICS YEAR TWO LEADERSHIP SERIES IMPLEMENTING STRATEGIES TO ENGAGE STUDENTS AND RAISE ACHIEVEMENT FEBRUARY 19, MARCH 26, MAY 21, 2014.
Department of Curriculum and Instruction Considerations for Choosing Mathematics Progress Monitoring Measures from K-12 Anne Foegen, Ph.D. Pursuing the.
Assessing Intelligence. Test Construction Standardization – defining the meaning of scores by comparing to a pretested “standard group”. Reliability –
Prepared by: Miss Samah Ishtieh
Assessing Intelligence
Lecture 5 Validity and Reliability
Questions What are the sources of error in measurement?
Session 3 – CAF as a first step towards a TQM strategy
Ministry of Defense of Georgia
Tests and Measurements: Reliability
ASSESSMENT OF STUDENT LEARNING
پرسشنامه کارگاه.
مركز مطالعات و توسعه آموزش دانشگاه علوم پزشكي كرمان
Reliability and Validity of Measurement
الاختبارات محكية المرجع بناء وتحليل (دراسة مقارنة )
USI College of Business Faculty Evaluation System
Oregon Teacher Standards & Practices Commission July 24, 2013
EPSY 5245 EPSY 5245 Michael C. Rodriguez
Measurement Concepts and scale evaluation
Closing – What Comes Next
Chapter 8 VALIDITY AND RELIABILITY
Kindergarten Scoring Practice Set – All Writing Tasks
Presentation transcript:

Reflections on the Migration of Medical Simulators from Training Tools to Assessment Tools Rich Shavelson Stanford Education Assessment Lab Invited Talk Simulation in Medical Education Seminar Stanford Hospital March 14, 2007

Overview Motivation for Talk Analogy and Implications: Medical Simulators and Job Performance Measurement Evaluation of Technical Quality of Simulator Assessments Open Discussion

Motivation For Talk Requests for advice on how to score performance collected with medical training simulators –SUMMIT—Surgical laparoscopic simulators Creating scores from data reported by simulator Defining reference groups Defining performance benchmarks Technical quality –Pediatric airway management simulator Separating training and assessment Developing a profile of performance Task sampling Technical quality Before Ideal & Possible Solution After Imple- menting Solution

Analogy and Implications: Medical Simulators and Job Performance Measurement A great deal of work has been done on job & education performance measurement that applies to medical simulators as assessment tools. My enduring interest

Analogy Continued Approaches to performance measurement –Construct definition –Task sampling definition –Rapprochement Cold reality Universe of Performance of Interest Universe of Possible Tasks for Assessment Universe of “Do-Able” Tasks Universe of “Do-Able” Task Formats Tasks & Scoring on Test

Evaluation Of Technical Quality Reliability –Classical test theory (“Cronbach’s Alpha”)— simple task sampling –Generalizability theory—task sampling –Item response theory—construct driven Validity –Construct –Content –Predictive

Thanks… Your Turn!