Measuring effectiveness - one foundation’s experience

Slides:



Advertisements
Similar presentations
What is District Wide Accreditation? Ensure Desired Results Improve Teaching & Learning Foster a Culture of Improvement A powerful systems approach to.
Advertisements

Local Control and Accountability Plan: Performance Based Budgeting California Association of School Business Officials.
The Periodic Review Report at the Community College: Opportunities for Collaborative Institutional Renewal Valarie Avalone, Director of Planning Dr. Michael.
Decision Making Tools for Strategic Planning 2014 Nonprofit Capacity Conference Margo Bailey, PhD April 21, 2014 Clarify your strategic plan hierarchy.
Improvement Service / Scottish Centre for Regeneration Project: Embedding an Outcomes Approach in Community Regeneration & Tackling Poverty Effectively.
Race to the Top Program Update January 30, State Funding 2.
Improving Government Effectiveness Tracy Gallo – State of Vermont June Sweeney - Office of the State Auditor.
Working Definition of Program Evaluation
Outcome Based Evaluation for Digital Library Projects and Services
+ Is your School's Instructional Program Ready for Common Core? Reach Institute for School Leadership.
2011 RP Conference: Growing Solutions. This session will give you: a sense of the bigger issues that are shaping our system an outline of what is happening.
Commissioning Self Analysis and Planning Exercise activity sheets.
Developing a Framework In Support of a Community of Practice in ABI Jason Newberry, Research Director Tanya Darisi, Senior Researcher
Kate Sandel & Brittany Anuszkiewicz The Finance Project November 16, 2007 Finding Funding for After School New Guides for California.
HLC Criterion Five Primer Thursday, Nov. 5, :40 – 11:40 a.m. Event Center.
Balanced Scorecard Process Writing Outcome Objectives and Measures.
Measuring effectiveness - one foundation’s experience M. Christine DeVita President The Wallace Foundation April 11, th Annual Conference The Council.
Orienting Agency Leads for Fund Mapping evidence2success Strategic Financing.
School Building Leader and School District Leader exam
Leading through a Time of Change
Strategic Planning – How it All Comes Together
Wayne County Foundation
Instructional Leadership for a Professional Learning Culture:
Planning for sustainable buildings
Measuring Impact Guide
Transforming the future of public health in Missouri
Success For Geneva’s Children Impact Study Danielle Shaw
THE SELF SUSTAINING NON-PROFIT Golden Lessons From the Development and Corporate Sectors 14th Eastern Africa Resource Mobilization Workshop Paper.
Driving Through the California Dashboard
Evaluation in Foundations
First 5 in 2017: Policy Outlook & System Approach
VASSP Conference – June 2016
Global Shelter Cluster Strategy Evaluation
Using Regional Groups to Effect Positive Change in HIV Care
Southern Regional Education Board Annual Leadership Forum
Introduction to Program Evaluation
Team Work and Accountability
Continuous Improvement through Accreditation AdvancED ESA Accreditation MAISA Conference January 27, 2016.
Collective Impact Fall 2017.
Achieving World-Class Cancer Outcomes: Taking the strategy forward May 2016 “People affected by cancer – those living with it and those supporting relatives.
Northwest Portland Area Indian Health Board Quarterly Meeting
Collective Impact Fall 2017.
9/16/2018 The ACT Government’s commitment to Performance and Accountability – the role of Evaluation Presentation to the Canberra Evaluation Forum Thursday,
One ODOT: Positioned for the Future
Research Program Strategic Plan
DESE Educator Evaluation System for Superintendents
RRI MONITORING AND EVALUATION
Summary.
Evaluating CLE Community Legal Education Masterclass
Planning for sustainable buildings
The Impact and Value of Instructional Coaching
Jean Scott & Logan Searcy July 22, MEGA
Purpose of This Document
Institutional Effectiveness Presented By Claudette H. Williams
Engaging Institutional Leadership
Administrator Evaluation Orientation
State of World’s Cash Report:
School Title I Stakeholder Meeting
Activity-Based Costing
Driving Through the California Dashboard
Strategy
Strategic Planning Final Plan Team Meeting
Changing the Game The Logic Model
Organizational Priorities
Reading Paper discussion – Week 4
Balanced Scorecard Process Writing Outcome Objectives and Measures
Tracie Wills Senior Commissioning Officer
Title I Document Training, Revision, Input Meeting
Using Data to Help Tell Your Story
Framing Grants for policy Research
Presentation transcript:

Measuring effectiveness - one foundation’s experience M. Christine DeVita President The Wallace Foundation April 11, 2005 56th Annual Conference The Council on Foundations

Evolution of The Wallace Foundation strategy Created 15 years ago through bequest of DeWitt and Lila Wallace Through 1999, made grants of nearly $1 billion to non-profits through 100+ different program Much accomplished -- but thought we could do better in fostering long-term improvements in areas we were working Board retreat in 1999 and agreement to address: Root causes rather than symptoms Seek improvements that outlasted grant dollars Seek impact beyond our grant dollars Contracted dozens of initiatives down to three: education leadership, arts participation, and out-of-school time learning Clarified our approach Support for innovation on the ground Develop evidence about what works and does not Share knowledge to inform the wider field

With new way of working, new challenge: measuring progress If it’s about catalyzing positive change in major public systems… Then to be effective, needed ongoing information about progress – for ourselves and our board See how the pieces – innovation, knowledge development, communication - fit together Because work is challenging and uncertain, ongoing feedback, revisions, mid-course corrections are crucial for success To have useful strategy discussions with board – a focus beyond individual grants – needed evidence of results

Driving the measurement goal – focus on good governance and effectiveness Good governance is the bedrock of effectiveness Investment performance Human resources Expenses Tougher challenge: What happens after money leaves the door Robert Wood Johnson’s scorecard was a model Decision: Aggregate progress of individual grantees within each focus area; produce report for annual board planning retreat in January

First try in 2003 – a focus on ‘what happened’ was a partial success The idea: Compare multiple progress measures against long-term desired results The how: Measured progress in three categories Grantee work Development of useful knowledge for the field Achievement of national benefits What we got: 23 pages of data on our focus area work – but little guidance on how to interpret Board comment: ‘This is very useful for internal management, but we need to know what it all means’

Second try in 2004 – not only ‘what happened’ but ‘what does it mean?’ Fine-tuned what we measured – added sustainability Working the plan Institutionalizing change Benefits to people Developing and spreading knowledge Asked one or two questions in each category to assess progress Answered each question with ‘core finding’

How do we make ‘core finding’ visible at a glance? Considered three options Traffic light approach: misleading Scores of 1-5: too definitive Eight-point progress gauges (little or none/modest/significant): much better Permits an assessment Allows comparisons Suggests relative, rather than exact, determinations Board comment: ‘A big improvement. The gauges are useful.’ Mixed/Modest Progress No/Unclear Progress Significant Progress

What it looks like in practice – working the plan Are core groups of grantees working the plan? Example Our work in 12 school districts to improve public education leadership The question: “Is leadership development improving?” Core finding: Increases in number and diversity of leaders getting leadership training and placed in high-need schools – but quality of training not systematically assessed Evidence: Participation in training activities up almost 20 percent over previous year, with 9,500 in total trained. About 70 percent of those trained are female, 40 percent people of color Gauge: Mixed or modest progress Mixed/Modest Progress No/Unclear Progress Significant Progress

What it looks like in practice – institutionalizing change Are sites making institutional changes? Example: Our work with arts organizations and state arts agencies to help them build participation The question: “Are grantees changing priorities to reflect participation-building?” Core finding: Non-Wallace spending on participation activities by our grantees declined slightly (2%). A minority of our state grantees adopted performance measures; many changed their strategic plans to focus on participation Evidence: Among arts organizations, 47 percent spent more on participation; 53 percent spent less. Among state arts agencies, 77 percent changed their strategic plans; 40 percent substantially increased focus on participation Gauge: Mixed or modest progress Mixed/Modest Progress No/Unclear Progress Significant Progress

What it looks like in practice – benefiting people Are people benefiting from our work? Example: Our work in arts participation The question: “Are more people from diverse demographic backgrounds attending arts performances or exhibits at the organizations we fund?” Core finding: Organizations that collect demographic data reported significant attendance gains among low-income and nonwhite participants from 2002 to 2003 Evidence: Median increases of 15% for low-income attendees (13 organizations reporting), and 11% for non-white participants (14 organizations reporting) Gauge: Significant progress Mixed/Modest Progress No/Unclear Progress Significant Progress

What it looks like in practice –producing useful knowledge Are we producing and sharing useful knowledge? Example: Our work in out-of-school time learning The question: “Are we producing knowledge that addresses key issues relevant to improving quality?” Core finding: Wallace-commissioned studies contribute to understanding demand and principles behind quality. No lessons yet on our work with cities to improve systems to support quality. Evidence: Publication of All Work and No Play, a Public Agenda survey Gauge: In between mixed and significant progress Mixed/Modest Progress No/Unclear Progress Significant Progress

Measures helped us determine action priorities Finding: We don’t know quality of leader training programs Action: Invest in assessment of quality Finding: Arts organizations are slightly cutting spending on participation-building Action: Invest in helping organizations build endowments or cash reserves for this purpose Finding: Practitioners find Wallace-commissioned studies helpful, but modest circulation of them Action: Invest in doing more to share knowledge

Lessons we’ve learned We needed to get the questions right ‘Absent a clear conception of effectiveness, foundations cannot assess (quantitatively or otherwise) whether or not they have achieved it, and risk adopting measures first and then adjusting their conceptions of success to fit the measures.’ Foundation Effectiveness, Francie Ostrower, The Urban Institute We needed to answer both ‘what happened?’ and ‘what did it mean?’ Board members found gauges useful to answer the latter Early indication: Helps board and staff get ‘big picture’ of how we are doing Early indication: Helps us plan more effectively

Challenges we face How easy or tough to grade? How do we account for barriers to action? How do we make the process efficient? How do we incorporate it into ongoing strategy development? How do we continue to engage board on effectiveness?

Measurement – a means to the end of greater effectiveness Provide input to strategy development The search for an appropriate measure starts with a clear description of what you want to achieve Use measurement of results to guide decisions about allocating funds between different initiatives Use measurement of results to guide decisions about what to invest within an initiative