Issues in Selecting Assessments for Measuring Outcomes for Young Children Issues in Selecting Assessments for Measuring Outcomes for Young Children Dale.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

Victorian Curriculum and Assessment Authority
Promoting Quality Child Outcomes Data Donna Spiker, Lauren Barton, Cornelia Taylor, & Kathleen Hebbeler ECO Center at SRI International Presented at: International.
CHANGING ROLES OF THE DIAGNOSTICIAN Consultants to being part of an Early Intervention Team.
Curriculum Based Evaluations Informed Decision Making Leads to Greater Student Achievement Margy Bailey 2006.
Comprehensive Curriculum Framework for Tiered Instruction: A Response to Intervention Model Sarah Jackson, M.Ed. Sandra Hess Robbins, M.Ed. Sanna Harjusola-Webb,
RtI in Early Childhood Lisa Kelly-Vance, UNO and Kristy Feden, Papillion- LaVista Schools.
Child and Family Outcomes Chapter 4: Using Assessment Tools in the Child Outcomes Measurement Process Entry & Exit Sample Information.
Grade 12 Subject Specific Ministry Training Sessions
Universal Screening and Progress Monitoring Nebraska Department of Education Response-to-Intervention Consortium.
What should be the basis of
performance INDICATORs performance APPRAISAL RUBRIC
The Current Status of States' Early Childhood Outcome Measurement Systems Kathy Hebbeler, SRI International Lynne Kahn, FPG Child Dev Inst October 17,
But What Does It All Mean? Key Concepts for Getting the Most Out of Your Assessments Emily Moiduddin.
Assessment in the early years © McLachlan, Edwards, Margrain & McLean 2013.
1 Implementation of the New Part C Eligibility Criteria Effective 7/1/2010.
The Child Outcomes Summary Competency Check (COS-CC) Amy Nicholas, Naomi Younggren, Siobhan Colgan & Kathi Gillaspy 2013 Improving Data, Improving Outcomes.
Wisconsin Departments of Health and Family Services (DHFS) And Public Instruction (DPI) OSEP Child Outcomes.
Preschool Outcomes Measurement System (POMS) Design and Implementation.
1 Topic 10: Role Of Program Assessment. Why Need Assessment? There are many reasons why children undergo assessments Desire to know how well children.
1 Birth to 3 Child Outcomes Maryland’s Approach to Converting Assessment Data to OSEP Outcome Categories August 28, 2007 Deborah Metzger
Classroom Assessments Checklists, Rating Scales, and Rubrics
1 Early Childhood and Accountability OSEP’s Project Director’s Meeting August 2006.
ND Early Childhood Outcomes Process Nancy Skorheim – ND Department of Public Instruction, Office of Special Education.
Aligning Child Outcome Measurement to Early Learning Standards NECTAC Outcomes Meeting Monday, August 27, 2007.
Progress Monitoring in Early Childhood: Generating a Discussion Judy Carta, Juniper Gardens, University of Kansas Nan Vendegna, Colorado Results Matter.
Approaches to Measuring Child Outcomes Kathy Hebbeler ECO at SRI International Prepared for the NECTAC National Meeting on Measuring Child and Family Outcomes,
Diagnostics Mathematics Assessments: Main Ideas  Now typically assess the knowledge and skill on the subsets of the 10 standards specified by the National.
Introduction Gathering Information Observation Interviewing Norm Referenced Tools Scores Administering Why, What, How Learning Check 5 Authentic Assessment.
Introduction Gathering Information Observation Interviewing Norm Referenced Tools Authentic Assessment Characteristics of Authentic Assessment – 7M’s Validity.
25 Industrial Park Road, Middletown, CT · (860) ctserc.org.
TESTS ARE US! 1)Accountability programs 2)Assessment & Testing 3)MCA MCA item samplers / test specifications / MCA reports.
Observation and Assessment in Early Childhood Feel free to chat with each other. We will start class at 9:00 PM ET! Seminar Two: Using Standardized Tests.
Early Childhood Special Education Part B, Section 619 Measurement of Preschool Outcomes-SPP Indicator #7 Training Sessions-2010.
Maryland’s Approach to Converting Preschool Outcomes Data to OSEP Reporting Categories Nancy M. Vorobey, M.Ed. Maryland State Department of Education
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Maryland’s Early Childhood Intervention and Special Education System of Services Birth through Five.
Early Childhood Outcomes Center New Tools in the Tool Box: What We Need in the Next Generation of Early Childhood Assessments Kathy Hebbeler ECO at SRI.
Session 2: Overview of the COS Process Child Outcomes Summary (COS) Process Training Module.
1 Early Childhood Assessment and Accountability: Creating a Meaningful System.
Candidate Assessment of Performance CAP The Evidence Binder.
Significant Developmental Delay Annual State Superintendent’s Conference on Special Education and Pupil Services October 20-21, 2015.
Candidate Assessment of Performance CAP The Evidence Binder.
Emerging Issues in Using Preschool and Kindergarten Assessment to Improve Learning Kathleen Hebbeler, DaSy/ECTA/IDC at SRI International Dee Gethman, Iowa.
Using Assessments to Monitor and Evaluate Student Progress 25 Industrial Park Road, Middletown, CT · (860) ctserc.org.
Johnson, J., Rahn, N., and Bricker, D., 2015 ECSE 672 Fall 2015 ACTIVITY-BASED APPROACH TO EARLY INTERVENTION.
Module 6: Using Tests for Assessment. Tests are important components to any comprehensive assessment system. Teachers, service providers, and specialists.
Observing and Assessing Young Children
School-Based Problem-Solving for Individuals (SBIT)
What do we know and what does it look like? Judith Carta & Virginia Buysse OSEP National Early Childhood Conference December 8, 2008 Washington DC.
WestEd.org Washington Private Schools RtI Conference Follow- up Webinar October 16, 2012 Silvia DeRuvo Pam McCabe WestEd Center for Prevention and Early.
Focus Questions What is assessment?
Facilitator: Angela Kapp Authentic Assessment Session 1 Session 1 Level 2 Minnesota Department of Human Services.
American Institutes for Research
Child Outcomes Summary Process April 26, 2017
Using Formative Assessment
Child Outcomes Summary (COS) Process Training Module
Diagnosis and Remediation of Reading Difficulties
Panel Discussion: What is being done by assessment tool developers and publishers to address OSEP reporting needs? Dale Walker, Charles Greenwood, University.
Kathy Hebbeler, ECO at SRI International AUCD Meeting Washington, DC
Lynne Kahn Kathy Hebbeler The Early Childhood Outcomes (ECO) Center
Building Capacity to Use Child Outcomes Data to Improve Systems and Practices 2018 DEC Conference.
Update from ECO: Possible Approaches to Measuring Outcomes
School Readiness and the Assessment of Children with Disabilities
Unit 7: Instructional Communication and Technology
Process for Developing Effective Functional, Routines-Based IFSPs
School Readiness and the Assessment of Children with Disabilities
Developing Effective Functional, Routines-Based IFSPs
Child Outcomes Summary (COS) Process Training Module
Process for Developing Effective, Functional, Routines-Based IFSPs
Presentation transcript:

Issues in Selecting Assessments for Measuring Outcomes for Young Children Issues in Selecting Assessments for Measuring Outcomes for Young Children Dale Walker & Kristie Pretti-Frontczak ECO Center and Kent State University Presentation at OSEP Early Childhood Conference Washington, DC, December 2005

2 Why Assessment? Gather information about skills and capabilities to make decisions about practice To determine eligibility for services To determine if a child is benefiting from services or if changes need to be made To measure development over time To document outcomes

3 Purpose of Assessments – It’s all about the question(s) you want to answer Screening – Is there a suspected delay? Does the child need further assessment? Eligibility determination – Is the child eligible for specialized services? Program planning – What content should be taught? How should content be taught? Progress monitoring – Are children making desired progress? Program evaluation/Accountability – Is the program achieving it intended outcomes and/or meeting required outcomes?

4 Assessment Options Norm-Referenced Criterion-Referenced Curriculum-Based Direct Observation Progress Monitoring Parent or Professional Report Any combination of assessments…

5 Norm-Referenced Pros/Cons Provides information on development in relation to others Already used for eligibility in many states Diagnosis of developmental delay Standardized procedures Do not inform intervention Information removed from context of child’s routines Usually not developed or validated with children with disabilities Do not meet many recommended practice standards May be difficult to administer or require specialized training

6 Norm-Referenced Assessment Table Table consists of a review of 18 norm- referenced assessments Information regarding each assessment is provided including: Publisher information Areas of development assessed Test norms provided Scores produced Age range covered

7 Criterion-Referenced Pros/Cons Measure child’s performance of specific objectives Direct link between assessment and intervention Provides information on children’s strengths and emerging skills Helps teams plan and meet individual children’s needs Meets recommended assessment practice standards Measures intra-child progress May be used to measure program effectiveness Requires agreement on criteria and standards Criteria must be clear and appropriate Usually does not show performance compared to other children Do not have standard administration procedures May not move child toward important goals Scores may not reflect increasing proficiency toward outcomes

8 Curriculum-Based Pros/Cons Provides link between assessment and curriculum Expectations based upon the curriculum and instruction Can be used to plan intervention Measure child’s current status on curriculum Evaluate program effects Often team based Meets DEC and NAEYC recommended standards Represents picture of the child’s performance May not have established reliability and validity May not have procedures for comparing child to a normal distribution Generally linked to a specific curriculum Often composed of milestones that may not be in order of importance

9 Curriculum-Based Assessment Rating Rubric Evaluates the quality of CBAs for use with young children Composed of 17 quality elements Used to guide teams in selecting appropriate CBAs

10 Sample of CBA Rubric ElementUnsatisfactory (0)Basic (1)Satisfactory (2)Excellent (3) Adaptable for Special Needs No consideration of special needs Limited consideration of special needs through the assessment process and instrument does not allow for additional accommodations or modifications for special needs Upfront considerations for special needs are not comprehensive, but assessment allows for some accommodations and/or modifications for special needs Considers and provides specific strategies and procedures for accommodating and/or modifying the assessment for special needs Aligns with Federal/State/Agency Standards and/or Outcomes Does not align with Federal/State/Agency Standards and/or Outcomes Aligns with less than half of the big ideas or concepts from Federal/State/Agency Standards and/or Outcomes Aligns with more than half of the big ideas or concepts from Federal/State/Agency Standards and/or Outcomes Aligns with a clear majority or all of the big ideas or concepts from Federal/State/Agency Standards and/or Outcomes

11 Progress Monitoring Pros/Cons Used to monitor ongoing progress toward important outcomes over time Compare to children of similar ages over time Repeatable measures for monitoring progress Standardized administration Standards for technical adequacy Efficient to administer May also be used as a screening tool Indicators of progress may be viewed as not being comprehensive Not used for eligibility determination May not provide specific skills to teach but indicators of important skills

12 Parent & Professional Report Pros/Cons High social validity Provides diverse perspective Important for informing intervention, program, IFSP/IEP Parents and professionals know the child, the environments in which they interact Collaboration requires time and effort to establish May not be reliable across time Does not permit comparison across children May include personal bias

13 Using Multiple Sources of Data or Single Source to Measure Outcomes? Pros and Cons Recommended practices Need to summarize information generated Ways data can be used beyond reporting OSEP outcomes

14 Using Data Beyond OSEP Reporting Good assessment data can be used to…. Reveal patterns regarding children’s strengths and emerging skills Develop functional and meaningful IFSPs/IEPs Inform program staff and families about strengths and weaknesses Guide the development of intervention Monitor children’s progress to inform intervention efforts Enhance collaboration Inform providers, programs, districts/parishes, regions, and states regarding important trends

15 Ongoing work and challenges… Existing assessment tools were not developed to measure the three outcomes ECO’s response: “Cross-walking” or mapping frequently used assessments to the outcomes Work with publishers and state staff to develop guidance for how to use assessment results to generate OSEP- requested data

16 Work with Publishers and Developers Finalizing crosswalks Alignment with OSEP outcomes How to determine what is “typical” performance Age-anchored benchmarks to measures How scores can be summarized using the ECO Summary Form Possible recalibration of scores in a way that maintains the integrity of different assessments Pilot studies with GSEG and interested states Data summary report forms that assist users with alignment of information from assessment to OSEP outcomes

17 Example of Developing a Validated Crosswalk First align On the face of it – which items appear to align/match which outcomes? Second validate Do experts agree? Check for internal consistency Third examine the sensitivity of the assessment in measuring child change

18 Example of Interpreting the evidence Standard scores Residual Change Scores Goal Attainment Scaling Number of objectives achieved/Percent objectives achieved Rate of Growth Item Response Theory Proportional Change Index Stoplight model

19 Interpreting the AEPS for Accountability First administration (near entry) Is the child above or below a cut off score? If above – considered to be developing typically If below – development is suspect Which level of the AEPS was administered? Child is less than three and Level I is used Child is less then three and Level II is used Child is older than three and Level I is used Child is older than three and Level II is used

20 Interpreting the AEPS for Accountability Second administration (near exit) Use cut off scores again Examine which level was used Look for changes in area percent scores changes in scoring notes changes in which level was administered

21 Sample Cutoff Scores LevelAge Intervals (months)Cutoff Score Birth to three Three to six

22 Questions?

For More Information see: For More Information see: For More Information see: For More Information see: