ASSESSING AND IMPROVING YOUTH PROGRAM QUALITY Nicole Yohalem October 2008.

Slides:



Advertisements
Similar presentations
Accountability for Quality: Policy Innovation in Out-of-school Time Programs November 2, 2007.
Advertisements

WV High Quality Standards for Schools
PORTFOLIO.
YOUTH PARTICIPATORY EVALUATION By: Nou Yang, Padah Vang, and Pao Thao.
Improving Quality Systemwide October 11, What is your role in afterschool?
Early Achievers Overview Starting Strong – August 15, 2012.
STRENGTHENING FAMILIES National Briefing, 2012.
Evaluating the Alternative Financing Program Geoff Smith Vice President Woodstock Institute March 18, 2008 WOODSTOCK INSTITUTE.
Linking Actions for Unmet Needs in Children’s Health
Karen L. Mapp, Ed.D. Deputy Superintendent, Boston Public Schools
Copyright © 2001 [Forum for Youth Investment]. All rights reserved. The Role of Out-of-School Programs as Blurring the Lines for Learning: The Role of.
Strengthening Families through Early Care and Education Columbia, SC October 14, 2006.
Common Ground One Approach, Many Adaptations Juanita Blount-Clark August, 2011.
School’s Out Washington Presentation to Skagit County Law & Justice Council Feb. 10, 2010 Janet Frieling, Network Director Shannon Ginn, Communications.
Family Resource Center Association January 2015 Quarterly Meeting.
Assessing & Improving Youth Program Quality Nicole Yohalem, Forum for Youth Investment Quality Youth Development Practice Webinar April 2, 2009.
Afterschool Programs: Expanding Learning, Reducing Achievement Gaps Afterschool Programs: Expanding Learning, Reducing Achievement Gaps Deborah Lowe Vandell.
What should be the basis of
performance INDICATORs performance APPRAISAL RUBRIC
Title I Needs Assessment/ Program Evaluation Title I Technical Assistance & Networking Session October 5, 2010.
Afterschool Programs: Fostering Youth Development AND Reducing Achievement Gaps Afterschool Programs: Fostering Youth Development AND Reducing Achievement.
TIMELESS LEARNING POLICY & PRACTICE. JD HOYE President National Academy Foundation.
Youth Development as a Public Health Policy: How to Make it Work Richard E. Kreipe, MD, FAAP, FSAM Professor of Pediatrics University of Rochester Leadership.
A joint venture between the Forum and High/Scope. Austin Youth Program Quality: A Snapshot Dr. Raphael Travis – Texas State University School of Social.
Systems Change 3 Integrated Jobs Strategy Policy Funding Outcome Data Capacity Development Innovation Leadership Values Collaboration Hall et al (2007)
Investing in Change: Funding Collective Impact
United Way of Greater Toledo - Framework for Education Priority community issue: Education – Prepare children to enter and graduate from school.
National Head Start Association Leadership Institute January 29, 2009 Presentation by Joan Lombardi, Ph.D. Early Childhood Development: At the dawn of.
United Way of Palm Beach County FY Applicant Orientation and Letter of Inquiry Training.
1 The Early Childhood Family Engagement Framework: Maryland’s Vision for Engaging Families with Young Children Jeffrey Capizzano President Maryland State.
Center for Youth Development and Policy Research National Leadership Summit on Improving Results Building a Youth Development Infrastructure Bonnie Politz.
Afterschool and STEM National Network of Statewide Afterschool Networks.
Creating a New Vision for Kentucky’s Youth Kentucky Youth Policy Assessment How can we Improve Services for Kentucky’s Youth? September 2005.
1 Understanding and Developing Child Welfare Practice Models Steven Preister, Associate Director National Child Welfare Resource Center for Organizational.
1 Adopting and Implementing a Shared Core Practice Framework A Briefing/Discussion Objectives: Provide a brief overview and context for: Practice Models.
Improving Program Quality in Central Texas. Agenda Overview Accomplishments Improvements Future Opportunities Break Assessors Methods Trainers.
Welcome Activity: Share Program Planning Ideas Use markers and sticky notes to post your ideas on the flip chart paper around the room! 1. Needs identified.
A New Adventure Begins for Program Quality! PROGRAM QUALITY ASSESSMENTS Presented by: Anne Reeder, Program Quality Assessment Coordinator DESE Grantee.
Hillsdale County Intermediate School District Oral Exit Report Quality Assurance Review Team Education Service Agency Accreditation ESA
Assessing Program Quality with the Autism Program Environment Rating Scale.
Ready by 21 Quality Counts Site Lead Meeting Highlights from Year One.
Evidence-based Evaluation for Afterschool Programs Denise Huang CRESST/UCLA 1/22/07.
An Insider’s Look at Improving Program Quality. The Ready by 21 Quality Initiative Forum for Youth Investment & The Robert Wood Johnson Foundation Austin.
Presented by Name, Title, Organization Phone Number: (xxx) INDIANA YOUTH WORKER CORE COMPETENCIES.
Building 21 st Century Skills with ICTs What does it mean? GEORGE SCHARFFENBERGER 6 September 2006.
Copyright © 2012 American Institutes for Research. All rights reserved. Oregon 21st Century Community Learning Centers Program Evaluation Mariel Sparr.
1 Quality Matters: Building Capacity and Investment in Youth Program Quality The Center for Youth Program Quality.
Council for Exceptional Children/Division of Early Childhood Conference October 2010 Kim Carlson, Asst. Director/619 Coordinator Ohio Department of Education.
Danielle Biselli & Mary Masla Mission To support, expand and advocate for quality out-of-school time programs and activities for children,
Do After-school Programs Affect Important Youth Outcomes? If So, Do We Know Why? Robert C. Granger, Ed.D. Remarks prepared for “Making a Difference in.
Your Presenters Melissa Connelly, Director, Regional Training Academy Coordination Project, CalSWEC Sylvia Deporto, Deputy Director, Family & Children’s.
Afterschool Programs…Keep kids safe, help working families and inspire learning. Education Policy Forum Breakfast Briefing November 1, 2007 Pittsburgh,
Bob Algozzine Rob Horner National PBIS Leadership Forum Chicago Hyatt Regency O’Hare October 8, /
Mountains and Plains Child Welfare Implementation Center Maria Scannapieco, Ph.D. Professor & Director Center for Child Welfare UTA SSW National Resource.
Section I: Bringing The Community Together Center for Community Outreach Key Components of Afterschool Programs.
Mountains and Plains Child Welfare Implementation Center Maria Scannapieco, Ph.D. Professor & Director Center for Child Welfare UTA SSW Steven Preister,
ANNOOR ISLAMIC SCHOOL AdvancEd Survey PURPOSE AND DIRECTION.
1 Early Childhood Assessment and Accountability: Creating a Meaningful System.
Colorado 9to25 Working together to improve outcomes for young people across Colorado.
Common Core Parenting: Best Practice Strategies to Support Student Success Core Components: Successful Models Patty Bunker National Director Parenting.
Developed by: July 15,  Mission: To connect family strengthening networks across California to promote quality practice, peer learning and mutual.
Promoting Family Economic Success in San Francisco.
1 A Multi Level Approach to Implementation of the National CLAS Standards: Theme 1 Governance, Leadership & Workforce P. Qasimah Boston, Dr.Ph Florida.
The PDA Center is funded by the US Department of Education Office of Special Education Programs Stories from the Field and from our Consumers Building.
Out of School Time Part III: Quality Connections to Quality Rating & Improvement Systems (QRIS) and School Age Settings Jennifer Harris, Arkansas State.
Agenda for Change Creating Stable Families Basic Needs Strategies and Guidelines.
Outcomes By the end of our sessions, participants will have…  an understanding of how VAL-ED is used as a data point in developing professional development.
Improving Quality Systemwide
Sel in ymca afterschool project results
Implementation Guide for Linking Adults to Opportunity
Presentation transcript:

ASSESSING AND IMPROVING YOUTH PROGRAM QUALITY Nicole Yohalem October 2008

© The Forum for Youth Investment 2008 Outline Intro discussion Making the case for investing in quality Where is the field? Defining Quality Assessing Quality Improving Quality

© The Forum for Youth Investment 2008 Features of Positive Developmental Settings National Research Council, 2002 Physical and Psychological Safety Appropriate Structure Supportive Relationships Opportunities to Belong Positive Social Norms Support for Efficacy and Mattering Opportunities for Skill-Building Integration of Family, School and Community efforts

© The Forum for Youth Investment 2008 “Communities should put in place some locally appropriate mechanism for monitoring the availability, accessibility and quality of programs…” - Community Programs to Promote Youth Development, 2002 National Research Council Recommendation

© The Forum for Youth Investment 2008 Discussion In your organizational context, which of these features do you consider strengths? Which represent areas of growth? Are you trying to measure any of these in your settings? Any common themes at your table?

© The Forum for Youth Investment 2008 Making the Case: Why focus on quality? 1. Because too few young people get the supports they need to thrive. According to the America’s Promise Alliance National Promises Survey, only 31% of 6-17 year olds have at least 4 of the 5 promises. 21% have 1 or none. 1. Safe Places 2. Caring Adults 3. Effective Education 4. Opportunities to Help Others 5. Healthy Start

© The Forum for Youth Investment 2008 Why focus on quality? 2. Because participating in OST programs can make a difference. Several empirical reviews of the effects of programs conducted over the past decade show that on average, programs have positive effects on social, emotional and academic outcomes. (Durlak & Weissberg, 2007; Bodilly & Beckett, 2005; Lauer et al., 2006).

© The Forum for Youth Investment 2008 Why focus on quality? 3. Because the quality of OST program matters. Durlak and Weissberg, programs reviewed grouped into 2 clusters based on SAFE criteria: (Sequenced, Active, Focused, Explicit) Programs that had the SAFE features showed positive effects on almost every outcome – school performance, social behavior, attitudes and beliefs. Programs that did not have the SAFE features showed no effect on any outcome. Other studies: Vandell, MARS, High/Scope’s YPQA Validation Study

© The Forum for Youth Investment 2008 Why focus on quality? 4. Because lots of children and youth spend time in these settings. Afterschool Alliance estimates 6.5 million children are in after-school programs. The parents of another 15.3 million say their child would participate were programs available. State of California moving toward universal access – elementary and middle.

© The Forum for Youth Investment 2008 Why focus on quality? 5. Because significant public and private dollars flow into these settings. The Finance Project estimated a $3.6 billion federal investment in after-school State and local funds increasingly important. California’s recent commitment of $550 million. Major private investments by national foundations (Wallace, Mott Atlantic, Clark, Robert Wood Johnson, W.T. Grant) and regional foundations (Nellie Mae, Skillman, William Penn, Colorado Trust).

© The Forum for Youth Investment 2008 Why focus on quality? 6. Because we know quality varies significantly within and across settings. All programs are not created equal (Durlak & Weissberg). Despite the variation, there are some patterns. Overall, programs do better on traditionally regulated things like safety than they do on higher- order things like engagement, interactions. Not all OST programs produce positive change. Funders can protect their investments by doing quality assurance.

© The Forum for Youth Investment 2008 Why focus on quality? 7. Because quality is measurable, and even more important – it is malleable. Palm Beach County, FL. Programs conduct assessments, sites develop improvement plans targeting specific areas, receive quality coaching, gains made in every area targeted by staff. American Camp Association. Based on youth surveys, leadership sets initial targets in partnership with youth and staff, develop an action plan, re-do survey. 17 out of the 23 camps had positive change in at least one area.

© The Forum for Youth Investment 2008 Why focus on quality? 8. Because measuring program quality is an efficient, affordable, and productive alternative (or complement) to measuring youth outcomes. There is more agreement in the field about the components of quality than about appropriate outcomes. Demonstrating impact is important, but doing it in such a way that individual programs can take credit for the change is extremely difficult, expensive and not necessarily instructive.

© The Forum for Youth Investment 2008 Why focus on quality? 1. Too few young people are getting the supports they need to thrive. 2. Participating in community programs can make a difference. 3. The quality of community programs matters. 4. Lots of children and youth spend time in these settings. 5. Significant public and private dollars flow into these settings. 6. We know quality varies significantly with and across settings. 7. Quality is measurable and malleable. 8. Measuring program quality can is an efficient, affordable, and productive alternative (or complement) to measuring youth outcomes.

© The Forum for Youth Investment 2008 Why focus on quality? Quality matters. Quality is measurable. Quality is malleable.

© The Forum for Youth Investment 2008 Ages Times of Day Outcome Areas (Cognitive, Physical, Social, Civic, Vocational…) ? ? ? Quality and Reach Count Morning... Night School After School At best, school only fills a portion of developmental space

© The Forum for Youth Investment 2008 Where is the field? Attention to and interest in quality has increased significantly over the past several years among key stakeholder groups in OST. There is consensus about what matters. Increasingly, evidence suggests our focus should be on the point of service. A lot of work is underway to refine assessment tools and develop systemic approaches to quality improvement. Washington is a site in the Ready by 21 Quality Counts initiative.

© The Forum for Youth Investment 2008 Measuring Youth Program Quality: A Guide to Assessment Tools Assessing Afterschool Program Practices Tool (APT) National Institute on Out-of-School Time and the MA Department of Education Out-of-School Time Observation Instrument (OST) Policy Studies Associates Program Observation Tool (POT) National Afterschool Association Program Quality Observation (PQO) Deborah Vandell and Kim Pierce Promising Practices Rating Scale (PPRS) WI Center for Education Research and Policy Studies Associates, Inc. Quality Assurance System (QAS) Foundations Inc. Program Quality Self-Assessment Tool (QSA) New York State Afterschool Network School-Age Care Environment Rating Scale (SACERS) Frank Porter Graham Child Development Center, UNC Youth Program Quality Assessment (YPQA) High/Scope Educational Research Foundation

© The Forum for Youth Investment 2008 Measuring Youth Program Quality: A Guide to Assessment Tools Purpose and History Content Structure and Methodology Technical Properties User Considerations Application in the Field

© The Forum for Youth Investment 2008 How is quality defined? There is a lot of similarity across definitions. Common elements include: Relationships Environment Engagement Social/Behavioral Norms Skill Building Opportunities Routine/Structure

© The Forum for Youth Investment 2008

How is quality measured? There are more differences in how quality is measured than in how it is defined. Why the differences in emphasis and approach? Tool purposes (regulatory, self-assessment) Program purposes (achievement, recreation) Developers’ perspectives & backgrounds (ECE, YD)

© The Forum for Youth Investment 2008 Differences in emphasis SACERSYPQA Social Processes/ Interactions 9 items14 items Resources (financial, human, material) 35 items16 items

© The Forum for Youth Investment 2008 Differences in approach Data collection methods Types of measures Technical properties

© The Forum for Youth Investment 2008 Data collection methods Type (observation, interview, questionnaire, document review) Target users (line staff, program leaders, youth, external observers) Intensity of data collection

© The Forum for Youth Investment 2008 Types of measures: high vs. low inference Program Observation Tool Staff are engaged with children Youth Program Quality Assessment During activities, staff generally smile, use friendly gestures, and make eye contact. Staff encourage all youth to try out new skills or attempt higher levels of performance. During activities, staff are almost always actively involved with youth (e.g. they provide directions, answer questions, work as partners or team members, check in with individuals or groups). Staff make use of frequent open-ended questions.

© The Forum for Youth Investment 2008 Types of measures: diagnostic vs. prescriptive Diagnostic – NY QSA A quality program provides participants with a variety of engagement strategies. Diagnostic and prescriptive - APT Youth are busy and engaged in conversation or activities. Youth appear relaxed and in control of themselves. Youth independently gather resources, materials or get information. Youth help select, lead or contribute to the running of the activity. Youth solve problems alone or in groups. When trying to solve a problem, youth try to identify the source, nature of the problem and/or try out potential solutions.

© The Forum for Youth Investment 2008 Technical properties

© The Forum for Youth Investment 2008 The YPQA is a good example of new class of tools that: Produce data that leads to real change in staff performance Provide continuity that is place-based (not silo-based) Link accountability policy with workforce development Offer a more efficient and effective use of resources than a sole focus on measuring child outcomes The YPQA: Assessing point of service quality

© The Forum for Youth Investment 2008 The YPQA: Assessing point of service q uality Plan Make choices Reflect Partner with adults Lead and mentor Be in small groups Experience belonging Engagement Reframing conflictEncouragement Skill building Active engagement Session flow Welcoming atmosphere Supportive Environment Interaction Safe Environment Healthy food and drinks Program space and furnitureEmergency procedures Psychological and emotional safety Physically safe environment

© The Forum for Youth Investment 2008 Safety Support Interaction Engage SAE System Accountability Environment PLC Professional Learning Community POS Point Of Service Quality in context: Multi-level systems

© The Forum for Youth Investment 2008 Youth PQA scores

© The Forum for Youth Investment 2008 What quality looks like on the ground “Positive Youth Development” 28% “Staff Centered” 39% “Low quality” 33% Sample of nearly 600 different youth workers

© The Forum for Youth Investment 2008 Approaches to Improving Quality Approaches differ along several dimensions: Nature of Agency Involvement (mandatory/voluntary) Level of Accountability (high stakes/low stakes) Reach (universal/targeted) Source of Expertise (internal/external capacity) Focus of Change (organizational issues/staff practice) Staff Level Targeted (targets leadership/line staff) Type of Data Collected (high/low inference; diagnostic/prescriptive) Support Strategy (one-on-one/group support)

© The Forum for Youth Investment 2008 Common lessons Quality assessment can advance multiple goals Data about their own practice is a powerful motivator for staff Common language helps pave the way for change Important to couple standards and assessment with tangible supports Building Quality Improvement Systems, 2007

© The Forum for Youth Investment 2008 mbus etroit Minneapolis Kentucky Iowa Oklahoma New York Rhode Island Austin Georgetown Divide Columbus Indianapolis Grand Rapids Nashville St. Louis Washington* West Palm Beach County Rochester Chicago Quality improvement systems using the YPQA YPQA is part of state and local quality improvement efforts: –Washington, New York, Iowa, Kentucky, Oklahoma (part of Ready by 21 Quality Counts) –Statewide 21st Century: Michigan, Maine, Minnesota, Rhode Island, New Mexico –Cities and Counties: Rochester, Syracuse, Detroit, Grand Rapids, Palm Beach, St. Louis, Nashville, Austin, Georgetown Divide CA, Indianapolis, Columbus IN, Chicago

© The Forum for Youth Investment 2008 Core components of a quality improvement system Self-assessment External assessment Action planning (with data) Training Coaching/advising

© The Forum for Youth Investment 2008 STEP 1 Decide to build system STEP 2a Program Self- assessment STEP 2b External assessment STEP 3 Plan for improve- ment STEP 4 Carry out plan IMPROVEMENT MONITORING ACCOUNTABILITY STEP 6 Re-assess and move forward! STEP 5 External assessment w/ criterion Components Self-assessment External assessment Action planning with data Training Coaching Example: Palm Beach, 60 sites county-wide; multi-year model

© The Forum for Youth Investment 2008 STEP 1 Decide to build system STEP 2a Program Self- assessment STEP 2b External assessment STEP 3 Plan for improve- ment STEP 4 Carry out plan IMPROVEMENT MONITORING ACCOUNTABILITY STEP 6 Re-assess and move forward! STEP 5 External assessment with criterion Components Self-assessment External Data Collection Action Planning (with Data) Training Coaching Example: Minnesota 115 sites statewide; multi-year model

© The Forum for Youth Investment 2008 What does the future hold? A QIS will gain more traction and more likely be effective and sustainable if the community/system also has solid information about: The overall program landscape (the full range of providers that could be engaged in quality improvement) The youth-serving workforce (who youth workers are, backgrounds and ambitions, turnover rates, full vs. part-time, etc.) Program participation Child and youth outcomes And if this information is part of an integrated information system

Measuring Youth Program Quality: A Guide to Assessment Tools Building Quality Improvement Systems Nicole Yohalem