Validity Evaluation NCSA Presentation NAAC/NCIEA GSEG CONSTORIA.

Slides:



Advertisements
Similar presentations
KATIE BUCKLEY, HARVARD UNIVERSITY SCOTT MARION, CENTER FOR ASSESSMENT NATIONAL CONFERENCE ON STUDENT ASSESSMENT (NCSA) NATIONAL HARBOR, MD JUNE 22, 2013.
Advertisements

Data Collection Strategies An Overview. How Do We Assess ?
1 Content-based Interpretations of Test Scores Michael Kane National Conference of Bar Examiners Maryland Assessment Research Center for Education Success.
Validity in Action: State Assessment Validity Evidence for Compliance with NCLB William D. Schafer, Joyce Wang, and Vivian Wang University of Maryland.
Evaluating the Validity of English Language Proficiency Assessments (EVEA): A Consortium Approach National Conference on Student Assessment Detroit, Michigan.
1 An Introduction to Validity Arguments for Alternate Assessments Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland.
Challenge Question: How would I like my performance on the job to be evaluated ? Self-Test Questions: 1.Why are some types of skills better assessed through.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.
Consistency/Reliability
Assessment Population and the Validity Evaluation
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments 1 Introduction to Comparability Inclusive Assessment Seminar.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
Classroom Assessment A Practical Guide for Educators by Craig A
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Accountability Assessment Parents & Community Preparing College, Career, & Culturally Ready Graduates Standards Support 1.
Chapter 2 Finding Ideas to Research. Generating Topics Translate ideas into valid and reliable ways of measuring them Collect evidence Unique Topics Innovative.
Teacher Certification Next Steps……. How certification works within your current practice Student Growth Criterion 3: Recognizing individual student learning.
The 5 E Instructional Model
Enhancing assessment capacity For teachers of Authority and Authority-registered subjects.
Thinking about assessment…
NEGOTIATING THE COMPLICATED TERRAIN OF TEACHER EVALUATION Isaak Aronson, Ph.D.
Becoming a Teacher Ninth Edition
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
ERIKA HALL CENTER FOR ASSESSMENT PRESENTATION AT THE 2014 NATIONAL CONFERENCE ON STUDENT ASSESSMENT NEW ORLEANS JUNE 25, 2014 The Role of a Theory of Action.
Analyzing COSF Data in Support of Validity Charles R. Greenwood July 25, 2007.
5 E’s Lesson Model.
Georgia Alternate Assessment Eighth Annual Maryland Conference October 2007 Melissa Fincher, Georgia Department of Education Claudia Flowers, UNCC.
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
Evaluating a Research Report
Classroom Assessment A Practical Guide for Educators by Craig A. Mertler Chapter 7 Portfolio Assessments.
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
CCSSO Criteria for High-Quality Assessments Technical Issues and Practical Application of Assessment Quality Criteria.
Assessment in Education Patricia O’Sullivan Office of Educational Development UAMS.
ATTRIBUTEDESCRIPTION Focal Knowledge, Skills, Abilities The primary knowledge / skills / abilities (KSAs) targeted by this design pattern. RationaleHow/why.
Alternative Assessment
Assessment Professional Learning Module 5: Making Consistent Judgements.
APRIL 16, 2013 A review of Chapters CHAPTER FOURTEEN The Role of Assessment VOCABULARY.
What makes great teaching? An overview of the CEM/Durham University/The Sutton Trust paper (published October 2014)......making... Capturing... promoting...
 Read through problems  Identify problems you think your team has the capacity and interest to solve  Prioritize the problems and indicate the.
Performance and Portfolio Assessment. Performance Assessment An assessment in which the teacher observes and makes a judgement about a student’s demonstration.
Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University.
ASSESSMENT, STANDARDS, & ACCOUNTABILITY. WHY IS ASSESSMENT IMPORTANT? Used to gather information and make decisions about student learning Formal tests.
Issues in Validity and Reliability Conducting Educational Research Chapter 4 Presented by: Vanessa Colón.
Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment.
Who are the Students in Alternate and Modified Achievement Standards Assessments? Jacqueline F. Kearns, Ed.D., NAAC Martha Thurlow, Ph.D., NCEO Elizabeth.
KVEC Presents PGES Observation Calibration Are You On Target?
Alternative Assessment Chapter 8 David Goh. Factors Increasing Awareness and Development of Alternative Assessment Educational reform movement Goals 2000,
What does exam validity really mean? Andrew Martin Purdue Pesticide Programs.
Assessment of Hands-on Science Including the Use of Portfolios By: Lori McEllin and Matt Shannon.
Chapter 6 - Standardized Measurement and Assessment
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
SCOTT MARION, CENTER FOR ASSESSMENT PRESENTATION AT CCSSO NCSA AS PART OF THE SYMPOSIUM ON: STUDENT GROWTH IN THE NON-TESTED SUBJECTS AND GRADES: OPTIONS.
GOALS 1. To understand RIBTS 9.5 as part of the process of assessing student learning. 2. To self-reflect on assessment information (e.g. pre-, post-,
Assessment Fall Quarter 2007.
Assessment of Learning 1
Test Validity.
Research on Using Observation Systems with Special Educators
Teaching and Educational Psychology
Validating Interim Assessments
Assessment Population and the Validity Evaluation
Mari Quenemoen Research Coordinator, NAAC
Assessment, Standards, & Accountability
Links for Academic Learning: Planning An Alignment Study
jot down your thoughts re:
Rubrics for evaluation
Alternative Assessment
jot down your thoughts re:
Assessment Fall Quarter 2007.
Presentation transcript:

Validity Evaluation NCSA Presentation NAAC/NCIEA GSEG CONSTORIA

Validation is “a lengthy, even endless process” (Cronbach, 1989, p.151)

NAAC/Center for Assessment. 4/30/08 3 Kane’s argument-based framework “…assumes that the proposed interpretations and uses will be explicitly stated as an argument, or network of inferences and supporting assumptions, leading from observations to the conclusions and decisions. Validation involves an appraisal of the coherence of this argument and of the plausibility of its inferences and assumptions (Kane, 2006, p. 17).”

Two Types of Arguments An interpretative argument specifies the proposed interpretations and uses of test results by laying out the network of inferences and assumptions leading from the observed performances to the conclusions and decisions based on the performances The validity argument provides an evaluation of the interpretative argument (Kane, 2006) NAAC/Center for Assessment. 4/30/08

The NAAC-GSEG Builds off of previous projects (NHEAI and first phase of NAAC) that focused on technical documentation Collecting and evaluating validity evidence was a significant challenge This project focuses on having states (5) create and evaluate validity arguments Our focus is NOT on single studies or even sets of studies, but on how the studies support or refute a validity argument NAAC/Center for Assessment. 4/30/08

Process Theory of Action Interpretive Argument Validity Evaluation

Theory of Action

Types of Studies Surveys – Teachers and Principals Skills, resources, supports – Scorer background – Consequential validity – Parent Surveys – Learner Characteristics Inventories Classroom Visit Protocols Scoring Observation Protocols Performance-level judgment assignment Focus group

Prioritizing the Studies Based on the Theory of Action… – Which studies will give us the best information to develop an interpretive argument? – How will we synthesize the study results into our interpretive argument?

Evaluating the Validity Argument Haertel (1999) reminded us that the individual pieces of evidence (typically presented in separate chapters of technical documents) do not make the assessment system valid or not, it is only by synthesizing this evidence in order to evaluate the interpretative argument can we judge the validity of the assessment program.

An approach for organizing synthesis for validity evaluation (Ryan & Marion) Example 1Example 2 Dimension of Validity Evidence Response processesGeneralization (reliability) Proposition/ Claim The rubric captures appropriately the knowledge and skills valued and assessed Scorers apply the rubric accurately and consistently when scoring the AA ‐ AAS Validity Study Questions What features and components of the GAA portfolio entries and scoring rubric do scorers attend to during the scoring process? Given training and support materials similar to those used by the state’s AA- MAS vendor, will scorers achieve an acceptable level of consistency and accuracy in their ratings? CriteriaWhat is acceptable or good enough evidentiary support? Typically r=.8? Evidence (Data) Collected Think-aloud protocols and interviews with a group of scorers (n = 14). Scores examined for consistency & accuracy for a set of trained scorers (w/out SWD expertise) ResultsSummary of resultsMixed results Claim and support? Is the claim for the proposition supported by the evidence? How strong is the support? Weak correlational results for inter- scorer agreement Potential alternative explanations? Do the results reveal other potential likely explanations? Scorers of AA-AAS must have content expertise to be effective scorers

Next Steps States are completing their studies and comparing their results with their interpretive arguments. They will then evaluate the study results interpretive argument comparisons to develop the validity evaluation.

Contact Information Jacqui Kearns, Ed.D. NAAC University of Kentucky Scott Marion NCIEA