© Regents of University of California 1 Functional Validity: Extending the Utility of State Assessments Eva L. Baker, Li Cai, Kilchan Choi, Ayesha Madni.

Slides:



Advertisements
Similar presentations
World-Class Instructional Design and Assessment
Advertisements

Understanding the ELA/Literacy Evidence Tables. The tables contain the Reading, Writing and Vocabulary Major claims and the evidences to be measured on.
Advances in the PARCC Mathematics Assessment August
California Educational Research Association Annual Meeting San Diego, CA November 18 – 19, 2010 Terry Vendlinski Greg Chung Girlie Delacruz Rebecca Buschang.
Common Core Standards and the Edmonds School District November 4, 2013.
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Eva L. Baker and Girlie Delacruz UCLA / CRESST Council of Chief State School Officers National Conference on Student Assessment Session: A Vital Goal:
Effective Intervention Using Data from the Qualitative Reading Inventory (QRI-5) Developed by the authors of the Qualitative Reading Inventory (QRI) -5,
Science Breakout New Teacher Meeting 6, Year 2 March 31, 2011.
WARNING: Making, carrying, or using overhead transparencies for every item in this training packet poses a significant physical and mental health hazard.
Overview of the CCSSO Criteria– Content Alignment in English Language Arts/Literacy Student Achievement Partners June 2014.
Explicit Direct Instruction Critical Elements. Teaching Grade Level Content  The higher the grade the greater the disparity  Test Scores go up when.
Types and Purposes of Assessment Mathematics Assessment and Intervention.
Making Sense of Assessments in HE Modules (Demystifying Module Specification) Jan Anderson University Teaching Fellow L&T Coordinator SSSL
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
1/16 CRESST/UCLA Alternative Assessment for English Language Learners Christy Kim Boscardin Barbara Jones Shannon Madsen Claire Nishimura Jae-Eun Park.
Principles of High Quality Assessment
The Art of Teaching Writing
Section VI: Comprehension Teaching Reading Sourcebook 2 nd edition.
ASSESSMENT IN EDUCATION ASSESSMENT IN EDUCATION. Reliability  Test-re-test, equivalent forms, internal consistency.  Test-re-test, equivalent forms,
English Language Development Assessment (ELDA) Background to ELDA for Test Coordinator and Administrator Training Mike Fast, AIR CCSSO/LEP-SCASS March.
NCCSAD Advisory Board1 Research Objective Two Alignment Methodologies Diane M. Browder, PhD Claudia Flowers, PhD University of North Carolina at Charlotte.
Launching the Common Core State Standards We need to prepare our students for 21 st Century Learning in an information age with technology innovations.
Evaluating Student Growth Looking at student works samples to evaluate for both CCSS- Math Content and Standards for Mathematical Practice.
NEXT GENERATION BALANCED ASSESSMENT SYSTEMS ALIGNED TO THE CCSS Stanley Rabinowitz, Ph.D. WestEd CORE Summer Design Institute June 19,
CCSSO-National Conference on Student Assessment 2013 National Harbor, MD– June 21, 2013 Girlie C. Delacruz, Eva L. Baker, Gregory K. W. K. Chung Solving.
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
Sample Middle School San Antonio Independent School District 2002–03 Name Principal.
Advances in the PARCC Mathematics Assessment August
Common Core State Standards (CCSS) September 12, 2012.
 Participants will teach Mathematics II or are responsible for the delivery of Mathematics II instruction  Participants attended Days 1, 2, and 3 of.
Forum - 1 Assessments for Learning: A Briefing on Performance-Based Assessments Eva L. Baker Director National Center for Research on Evaluation, Standards,
© 2012 Board of Regents of the University of Wisconsin System, on behalf of the WIDA Consortium Introduction to the WIDA English Language Development.
The present publication was developed under grant X from the U.S. Department of Education, Office of Special Education Programs. The views.
THE DANIELSON FRAMEWORK. LEARNING TARGET I will be be able to identify to others the value of the classroom teacher, the Domains of the Danielson framework.
Part II: Who are the students?Page 1 Part II: Who are the Students who take Alternate Assessments on Alternate Achievement Standards? Articulating the.
1 Race to the Top Assessment Program General & Technical Assessment Discussion Jeffrey Nellhaus Deputy Commissioner January 20, 2010.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
So Much Data – Where Do I Start? Assessment & Accountability Conference 2008 Session #18.
Assessing The Next Generation Science Standards on Multiple Scales Dr. Christyan Mitchell 2011 Council of State Science Supervisors (CSSS) Annual Conference.
Baker ONR/NETC July 03 v.4  2003 Regents of the University of California ONR/NETC Planning Meeting 18 July, 2003 UCLA/CRESST, Los Angeles, CA ONR Advanced.
ONR/NSF Technology Assessment of Web-Based Learning, v3 © Regents of the University of California 6 February 2003 ONR/NSF Technology Assessment of Web-Based.
1 Alignment of Alternate Assessments to Grade-level Content Standards Brian Gong National Center for the Improvement of Educational Assessment Claudia.
Sample Elementary San Antonio Independent School District 2002–03 Name Principal.
Target -Method Match Selecting The Right Assessment.
COUNCIL OF CHIEF STATE SCHOOL OFFICERS (CCSSO) & NATIONAL GOVERNORS ASSOCIATION CENTER FOR BEST PRACTICES (NGA CENTER) JUNE 2010.
Student report cards Presentation for Primary School Staff 2007.
Smarter Balanced Interim Assessment System. Session Overview What are the interim assessments? How to access? How to score? Using the THSS and the scoring.
Bridge Year (Interim Adoption) Instructional Materials Criteria Facilitator:
Summary of Assessments By the Big Island Team: (Sherry, Alan, John, Bess) CCSS SBAC PARCC AP CCSSO.
You Can’t Afford to be Late!
State Board of Education Presentation Jeffrey Hauger, Ed.D. Peggy McDonald, Ed.D. Elizabeth Celentano, M.Ed. January 11, 2016 NEW JERSEY DYNAMIC LEARNING.
Measuring College and Career Readiness 2015 PARCC RESULTS: YEAR ONE EDGEWATER SCHOOL DISTRICT ELEANOR VAN GELDER SCHOOL.
LITERACY-BASED DISTRICT-WIDE PROFESSIONAL DEVELOPMENT Aiken County Public School District January 15, 2016 LEADERS IN LITERACY CONFERENCE.
1 Science, Learning, and Assessment: (Eats, Shoots, and Leaves) Choices for Comprehensive Assessment Design Eva L. Baker UCLA Graduate School of Education.
Maxson Bilingual/ESL Program Type of Bilingual Program Developmental (Content area taught in Spanish) -Language Arts -Math -Science -Social Studies Type.
Georgia Milestone Assessments Bryant Elementary School 1"Where Scholars Are Prepared to Be LEADERS"
Building an Interim Assessment System: A Workbook for School Districts CCSSO National Conference on Student Assessment Detroit, MI June 22, 2010.
Building Effective Content Literacy Tasks. The Cycle of Assessment Teach: Does the instruction and the tasks align to the identified learning target(s)?
Welcome Parents! FCAT Information Session. O Next Generation Sunshine State Standards O Released Test Items O Sample Test.
Colorado Academic Standards Colorado English Language Proficiency (CELP) Standards There are now five English language development standards: Standard.
Donna Lipscomb EDU 695 MAED Capstone Common Core Presentation INSTRUCTOR KYGER MAY 21, 2015.
Research Questions  What is the nature of the distribution of assignment quality dimensions of rigor, knowledge construction, and relevance in Math and.
Dr Anie Attan 26 April 2017 Language Academy UTMJB
Pre-Referral to Special Education: Considerations
M-Step Overview, Practice and Instructional Implications
Network Team Institute July 8-12, 2013
Claudia Flowers, Diane Browder, & Shawnee Wakeman UNC Charlotte
Deputy Commissioner Jeff Wulfson Associate Commissioner Michol Stapel
Claudia Flowers, Diane Browder, & Shawnee Wakeman UNC Charlotte
Presentation transcript:

© Regents of University of California 1 Functional Validity: Extending the Utility of State Assessments Eva L. Baker, Li Cai, Kilchan Choi, Ayesha Madni UCLA/CRESST Comparing Expectations for Validity Models and for New Assessments: Goals, Approaches, Feasibility, and Impact Council of Chief State School Officers (CCSSO) 2015 National Conference on Student Assessment San Diego, California – June 24, 2015

© Regents of University of California 2 So What’s New? Opting Out Salient target Evidence of benefit Displaced anger Transparency

© Regents of University of California 3 Transparency: Expectations Clear and Sensible? Better test transparency and utility for public Specificity in the right places Support student learning and persistence

© Regents of University of California 4 Today: Feature Analysis Argue that tests for “summative” purposes can contribute to transparency of findings to improve learning By conducting qualitative and quantitative analyses of tests (and interventions) the veil of obscurity—and what to teach—can be lifted FA key element of data-mining

© Regents of University of California 5 Features for Analysis and Design of Assessments

© Regents of University of California 6 How It Works Rate components of items/tasks Low inference features Features recombined in tasks, items (game levels, episodes) Meta-tagged in data Performance summaries across individual or clusters of features Criteria: Significant difficulty, growth, or complexity

© Regents of University of California 7 Sample CRESST Features: Content, Cognition, Task, Linguistics Knowledge—mapped to standards and prerequisites – Content—topics, memory, concepts, procedures, systems – Representations Cognitive requirements and skills – Problem solving components – Communication, inferencing – Pattern detection, situation awareness

© Regents of University of California 8 Task Features Surface requirements – Format – Stimulus content, prompts, resources, representations – Game mechanic or interaction engine – Affordances, accessibility, accommodations – Team work requirements – Narrative or scenario content and structure Response Requirements – Answer formats – Criteria or scoring rules – Actions or number of types and steps in a response – Essay elements or particular demands

© Regents of University of California 9 Linguistic Features Discourse – Complexity or number of ideas in passage or directions – Length – Literal or inferential comprehension – Academic structure, domain-dependent or independent Syntax – Sentence patterns, type and variation – Sentence length – Context cues Word choice – Academic vocabulary-specific domain – Academic language, type, density

© Regents of University of California 10 Problem Solving Constraints Single: Increase Vector’s speed to reach stars by reducing amount of friction. Multiple: Increase Vector’s speed to reach stars but not too fast to avoid hitting dynamite.

© Regents of University of California 11 State Assessment Study - 1 Purpose: To predict performance from three years of standards use and attribute results – Rated features of content, cognition, linguistics, and tasks with high consistency – Tagged every test item in English Language Arts (ELA) and math for grades 3 & 4 and 7 & 8 for years 2011, 2012, and 2013 by feature – Features accounted for on average 50% of variance on item difficulty

© Regents of University of California 12 Assessment Study - 2 Math Grades 4, 8, 11 Previous features augmented by results of student think-alouds A total of 70 features identified and tagged on a sample of math items

© Regents of University of California 13 Assessment Study - 2 Findings 4th grade: 16 features were significantly related to difficulty, 10 harder, 6 easier 8th grade: 10 features significantly related to difficulty, 6 harder, 4 easier 11th grade: 12 features related to difficulty, 7 harder, 5 easier

© Regents of University of California 14 Feature Relationships Features across grades – Cognitive load – Representation type – Constructed or multiple responses – Guidance – Linguistics Features across ELA and math – Linguistics amount Able to predict item difficulty by features

© Regents of University of California 15 Current R&D Continuing FA of state level assessments, refining definitions, protocols, and training Sub-group and feature interactions FA of interventions—PBS learning games and videos, classroom instructional assignments Linking features of interventions and assessments to predict performance Developing two ways of automated feature extraction Designing assessments and games using features Engaging in FA validity studies across projects Looking for partners

© Regents of University of California 16 Summary Feature analysis may make “summative” results useful for improvement Multiple purposes for tests Development implications for tests and for designing and predicting effects of interventions

Copyright © 2014 The Regents of the University of California. Do Not Distribute Eva L. Baker

© Regents of University of California 18 Back up slides Back Up Slides

© Regents of University of California 19 State Assessment Functional Validity Data to determine year-to-year cohort performance changes – instructional sensitivity? Summarized across specified features significantly related to high and low difficulty Resulting feature sets accounted on average for 50% of variation of performance If confirmed by instructional studies, findings may guide teachers and professional development to improve test performance using invariant Guide procurement for re-designed specifications Note: Cai, Baker, Choi, Buschang, 2014; Baker, Cai, Choi, 2014; Choi, Madni, 2015

© Regents of University of California 20 How It Is Done – Feature Parsing Elements are defined and rated which comprise test items and tasks or learning requirements, e.g., linguistics, content elements, detailed cognitive processes Each item is re-rated by pairs of trained staff for each feature. More granular and operational level of analysis than many currently used approaches Features tagged to items in data

© Regents of University of California 21 Purposes of Assessments Beyond accountability Policy linking accountability and improvement Accountability analyses interfere with guidance supporting teaching and learning Can improvement of learning become a useful function of large-scale tests? Feature analysis of item and test properties can yield useful instructional information

© Regents of University of California 22 Mapping Features: Ontologies: Networks of Relationships SEL Problem Solving Content