Systematic Naturalistic Inquiry: Toward a Science of Performance Improvement (aka improvement research) Anthony S. Bryk Carnegie Foundation for the Advancement.

Slides:



Advertisements
Similar presentations
Intelligence Step 5 - Capacity Analysis Capacity Analysis Without capacity, the most innovative and brilliant interventions will not be implemented, wont.
Advertisements

Building Effective Leadership Teams: A Practitioner’s Look
LINDSAY CLARE MATSUMURA HELEN GARNIER BRIAN JUNKER LAUREN RESNICK DONNA DIPRIMA BICKEL June 30, 2010 Institute of Educational Sciences Conference Evidence.
Download an electronic copy of the slides and a handout here: projectgladstudy.educationnorthwest.org.
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
Spending Public Money Wisely Scaling-Up Educational Interventions Barbara Schneider John A. Hannah University Distinguished Professor College of Education.
1 Reading First Internal Evaluation Leadership Tuesday 2/3/03 Scott K. Baker Barbara Gunn Pacific Institutes for Research University of Oregon Portland,
Final Session: Whole Group Please sit together with the other associates from your district (as much as possible) April 27, 2005.
5E INQUIRY LEARNING CYCLE. EngageExplore ExplainElaborate Evaluate.
Leadership within SW-PBS: Following the Blueprints for Success Tim Lewis, Ph.D. University of Missouri OSEP Center on Positive Behavioral Intervention.
The Impact of Literacy Coaching on Teachers’ Value- Added to Student Learning in Literacy Collaborative Gina Biancarosa, University of Oregon Anthony S.
Professional Growth= Teacher Growth
+ Hybrid Roles in Your School If not now, then when?
Sharon Walpole University of Delaware Michael C. McKenna University of Virginia Literacy Coaches in Action: Strategies for Crafting Building- Level Support.
Developing School-Based Systems of Support: Ohio’s Integrated Systems Model Y.S.U. March 30, 2006.
Aligning Observations and In-service Professional Development Prek-3 rd Jason Downer May 9, 2012.
Professional development for mainstream teachers of ELLs: Project GLAD ® and Beyond Theresa Deussen March 10, 2014.
“We will lead the nation in improving student achievement.” CLASS Keys TM Module 1: Content and Structure Spring 2010 Teacher and Leader Quality Education.
“Knowing Revisited” And that’s how we can move toward really knowing something: Richard Feynman on the Scientific Method.
District Workforce Module Preview This PowerPoint provides a sample of the District Workforce Module PowerPoint. The actual Overview PowerPoint is 62 slides.
Striving to Link Teacher and Student Outcomes: Results from an Analysis of Whole-school Interventions Kelly Feighan, Elena Kirtcheva, and Eric Kucharik.
Evaluating the Vermont Mathematics Initiative (VMI) in a Value Added Context H. ‘Bud’ Meyers, Ph.D. College of Education and Social Services University.
Coaching in Literacy Collaborative and Its Effects on Teachers and Students Gina Biancarosa, University of Oregon Anthony S. Bryk, Carnegie Foundation.
What Was Learned from a Second Year of Implementation IES Research Conference Washington, DC June 8, 2009 William Corrin, Senior Research Associate MDRC.
Full Implementation of the Common Core. Last Meeting Performance Tasks Smarter Balanced Assessment Consortium Upcoming Accountability Measure Strong teaching.
Setting purposeful goals Douglas County Schools July 2011.
The Instructional Decision-Making Process 1 hour presentation.
TAP TAP Basics (Preparing for Success in a TAP School) [PSTS]
Distributed Leadership for Mathematics Bringing Together District, School, & University Leadership to Support Highly Qualified Teachers University of Wisconsin-Milwaukee.
KATEWINTEREVALUATION.com Education Research 101 A Beginner’s Guide for S STEM Principal Investigators.
Cohort 5 Middle/Jr. High School Data Review and Action Planning: Schoolwide Reading Spring,
1. Housekeeping Items June 8 th and 9 th put on calendar for 2 nd round of Iowa Core ***Shenandoah participants*** Module 6 training on March 24 th will.
LEAP Conference July 2015 Helen Timperley. Professional development to build capability is fundamental to whole school improvement processes BUT...
How Principals Build Organizational Capacity to Improve Student Learning Presentation to the Illinois School Leadership Advisory Council January 2015 By:
Teresa K. Todd EDAD 684 School Finance/Ethics March 23, 2011.
Eloise Forster, Ed.D. Foundation for Educational Administration (FEA)
What makes great teaching? An overview of the CEM/Durham University/The Sutton Trust paper (published October 2014)......making... Capturing... promoting...
Why is research important Propose theories Test theories Increase understanding Improve teaching and learning.
Evidence-Based Medicine Presentation [Insert your name here] [Insert your designation here] [Insert your institutional affiliation here] Department of.
Standards-Based Curricula Dr. Dennis S. Kubasko, Jr. NC Teach Summer 2009.
UDL Coaching: What, Why, and How Patti Ralabate Director of Implementation March 5, 2014.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Math and Science Partnership Program Approaches to State Longitudinal Evaluation March 21, 2011 San Francisco MSP Regional Meeting Patty O’Driscoll Public.
Competency 010 Presented by Elizabeth Castillo. Competency 010 The Master Technology Teacher knows how to provide professional development and support.
OSEP Project Directors’ Conference Washington, DC July 21, 2008 Tools for Bridging the Research to Practice Gap Mary Wagner, Ph.D. SRI International.
Applying impact evaluation tools A hypothetical fertilizer project.
TEACHER EFFECTIVENESS INITIATIVE VALUE-ADDED TRAINING Value-Added Research Center (VARC)
The Practical Aspects of Doing Research An Giang University June, 2004 Dennis Berg, Ph.D.
Growth and Goals Measures of Academic Progress ® (MAP ® ) Measures of Academic Progress, MAP, and DesCartes: A Continuum of Learning are registered trademarks.
DVAS Training Find out how Battelle for Kids can help Presentation Outcomes Learn rationale for value-added progress measures Receive conceptual.
Insights About the Grade 1 Learning Expectations in the CCSS Topic #1: The meaning of the equal sign Presented by the Office of Curriculum, Instruction.
Developing a Framework In Support of a Community of Practice in ABI Jason Newberry, Research Director Tanya Darisi, Senior Researcher
FASA Middle School Principal ’ s Leadership Academy Don Griesheimer Laura Hassler Lang July 22, 2007.
Changes in Professional licensure Teacher evaluation system Training at Coastal Carolina University.
Design of Clinical Research Studies ASAP Session by: Robert McCarter, ScD Dir. Biostatistics and Informatics, CNMC
“CEO Louis Gerstner added $40 billion to IBM’s stock market value.” “By himself?” Linking Educational Leadership to Student Achievement 2008 IES Conference.
Greenbush. An informed citizen possesses the knowledge needed to understand contemporary political, economic, and social issues. A thoughtful citizen.
A TAP Story: A. A. Nelson Elementary School Jacqueline Smith, Principal A.A. Nelson Elementary School TAP Leadership Team Teddy Broussard, State TAP Director.
Principal Student Achievement Meeting PLC Visioning and Beyond.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
What is Learning-Focused?
Using PLCs to Build Expertise Community of Practice October 9, 2013 Tammy Bresnahan & Tammy Ferguson.
Evaluating Growth Patterns. Setting the Stage  Welcome/Introductions  Structure for the day  Materials review R A M Materials Reports Activity M A.
STRONG FAMILY AND COMMUNITY ENGAGEMENT IS CENTRAL TO EFFECTIVE SCHOOL REFORM Jan Patterson and Ann Bliss Smarter Schools National Partnerships Key Reform.
Laying the Foundation A Discussion on Moving Fidelity of Implementation from Compliance to Capacity Building Carol K. McElvain American Institutes for.
Data-Driven Instructional Leadership
Align Combine Design.
5E Inquiry learning cycle
Scientific Methods Science in Practice.
Considering Fidelity as an Element Within Scale Up Initiatives Validation of a Multi-Phase Scale Up Design for a Knowledge-Based Intervention in Science.
Presentation transcript:

Systematic Naturalistic Inquiry: Toward a Science of Performance Improvement (aka improvement research) Anthony S. Bryk Carnegie Foundation for the Advancement of Teaching Society for Research on Educational Effectiveness, March 2010

I. Revisiting a 30+ year old argument Is design really the answer? The randomized treatment control paradigm as the gold standard circa 1975 Takes me back to the spring of 1978: “evaluating program impact: a time to cast away stones, a time to gather stones together” And, is this really the right question?

II. What Information Does an RCT Actually Provide? –Two marginal distributions Y T and Y C : the distributions of outcomes under the treatment and control conditions. –Provides answers to questions that can be addressed in term of observed differences in these two marginal distributions.

Evidentiary Limits of the Treatment- Control Group Paradigm Suppose now that we define a treatment effect for individual i as α i. We can estimate the mean treatment effect, μ α. But, interestingly we cannot estimate the median effect or any percentile points in the α i distribution.

Evidentiary Limits (continued) Nor can we assess any linkages between α i and how these effects might be changing over time, or depend on individual and context characteristics. –To accomplish the latter, we need to know about the treatment effect distribution conjoint with multivariate data on individual and program characteristics.

Evidentiary Limits (continued) Of course we can add a limited number of factors into the design and estimate these interaction effects. So we can do something on a limited scale within the T/C paradigm –But we need to know the factors in advance –And they have to be small in number –Pushing the envelop here would be time- consuming, expensive and cumbersome

My conclusions back then We need a different methodology for learning about programs and the multiple factors that may affect their outcomes –An accumulating evidence strategy (Light and Smith) from multiple efforts at systematic inquiry over time –Needs to be dynamic in design—as we learn from practice we are changing it –A system orientation— “elements standing in strong interaction.” pause

The Paradox of Anti-depressant- Induced Suicidiality (H.I. Weisberg, V.C. Hayden, V.P. Pontes (2009) Clinical Trials. Vol 6.No. 2, ) Key conclusions: –When the causal effect of an intervention varies across individuals the threat to validity can be serious. –RCTs should not automatically be considered definitive, especially when the results conflict with those of observational studies. –Not only the magnitude but even the direction of the population causal effect may be erroneous.

III. So a New Directions 2010: Basic Principles Returning to this idea of a prospective accumulating evidence strategy –Simplest version: the multi-site trial vs. cluster randomized trial. –Extend this idea out to all three facets – contexts, teachers, and students.

Basic Principles Anchored in a working theory about advancing improvements reliably at scale - Assume a systems perspectives: interventions as operationally defined in strong interactions with the specific people who take it up and the contexts in which they work. –Gathering and using empirical evidence about such phenomena should be the organizing goal.

Basic Principles (continued) Accelerated longitudinal design + a value added analytic model. –Counterfactual comes from a baseline comparison. –In principle we have some evidence about variable effects attached to individuals, their teachers and their context. –Any individual piece not very precise but if we have enough cases there is power to see many signals.

Basic Principles (continued) A key internal validity concern – a coterminous intervention to worry about. But we also now have an evidentiary resource not typically found in RCT –a capacity to examine questions of replicability over many different contexts of intervention. –This is the generalizability evidence that relates directly to our reliability consideration. –Can we make this happen with any reliability over many different situation?

What makes it naturalistic? Easily engaged in practice. Could be routinely done. Could imagine gathering such data at large scale. Immediacy of evidence – possibility of learning as you go. And as it will turn out, actually moot (opportunistic) on the question of an appropriate design analysis paradigm

A recently completed study of the efficacy of Literacy Collaborative Professional Development Co-contributor: Gina Biancarosa University of Oregon Detailing the causal cascade from the intentional design of professional education through changes in instructional practice and then on through to improvements in student learning gains over time. IV. Elaborate through an Example

Setting the Context: Typical District Approach to a Coaching Initiative Credit to: A Framework for Effective Management of School System Performance. Lauren Resnick, Mary Besterfield-Sacre, Matthew Mehalik, Jennifer Zoltners Sherer and Erica Halverson.

And then voila! (aka the zone of wishful thinking!)

Peering inside the “Black Box”: the actual work of coaches

Data for Performance Improvement What do coaches need to know and be able to do and how do we know if they can? Quality of the trust dependency/ relationship How do coaches actually spend their time? Quality of coach-teacher trust social resources for improvement Who is being coached on What topics? What about the Individual teacher might affect These social exchanges?

Data for Performance Improvement Teacher practice development Evidence of teacher learning ?

Filling out the account: an information system to support instructional improvement Coaching Logs Coaching performance assessments Surveys of coach principal trust: respect, regard, competence and integrity Surveys of teacher-coach trust and school-based professional community Teacher practice development Observational evidence of teacher learning and practice Coaching logs: the who and what of PD as delivered and what’s next?

Individual Teacher Classroom Literacy Practice School-wide support for teacher learning Formal School Structure Informal Organization * principal leadership * coach quality/role relationship * resource allocations (time) * school size * Work relations among teachers * Influence of informal leaders *professional norms Joined in a Working Theory of Practice Improvement Background Willingness to engage innovation –Experiment with new practices in the classroom Expertise –Prior experiences in comprehensive literacy teaching (ZPD) LC Intervention: amount, quality and content Of PD Impact on Student learning It is hard to improve what you do not really understand.

Linked to evidence about variability in effects on student learning associated with teachers and schools Assessing (even crudely) the value added to learning associated with individual classrooms and schools and investigating what might be driving observed variability in these effects.

Accelerated Cohort design 6 cohorts studied over 4 years Year of Study First YearSecond YearThird YearFourth Year FallSpringFallSpringFallSpringFallSpring K CCDDEEFF 1 BBCCDDEE 2 AABBCCDD Grade Training year Year 1 of implementation Year 2 of implementation Year 3 of implementation

Y tijk time Latent individual initial status,π 0i Latent individual Growth rate,π 1i v 1jk v 2jk v 3jk v 4jk Basic value added model ŷ 0ijk = π 0i ŷ lijk =π 0i + π li + v 1jk ŷ 2jjk =π 0i + 2π li + v ljk + v 2jk ŷ 3jk =π 0i + 3π li + v ljk + v 2jk + v 3jk ŷ 4jk =π 0i + 4π li + v ljk + v 2jk + v 3jk +v 4jk Gain from year t -1 to t = π li + υ tjk Note: v jk may vary over time as well. The Logic of a Value-Added Model for Assessing Impact on Student Learning v tjk,value-added at time t Observed growth data

Hierarchical Crossed Value-added Effects Model Individual growth parameters overall value- added effects teacher-level school-level value-added effects

Value-added effects by year Year 1Year 2Year 3 Average value-added (overall) Performance improvement 16%28%32% Effect size Ave. student learning growth is 1.02 per academic year

.16 = Year 1 mean value-added.28 = Year 2 mean value-added.33 = Year 3 mean value-added 1.02 = 100% value-added 0 = Baseline growth rate (no value-added)

.16 = Year 1 mean value-added.28 = Year 2 mean value-added.33 = Year 3 mean value-added 1.02 = 100% value-added 0 = Baseline growth rate (no value-added)

School 95% plausible value- added range ±.23 Variability in school value-added, year 1 Average student gain per academic year No effect Year 1 mean effect

School 95% plausible value- added range ±.23±.28 Variability in school value-added, year 2 Average student gain per academic year No effect Year 1 mean effect Year 2 mean effect

School 95% plausible value- added range ±.23±.28±.37 Variability in school value-added, year 3 Average student gain per academic year No effect Year 1 mean effect Year 2 mean effect Year 3 mean effect

School 95% plausible value- added range ±.23±.28±.37 Variability in school value-added, year 3 Average student gain per academic year No effect Year 1 mean effect Year 2 mean effect Year 3 mean effect

Teacher 95% plausible value- added range ±.51 Variability in teacher value-added within schools, year 1 Average student gain per academic year No effect

Teacher 95% plausible value- added range ±.51±.71 Variability in teacher value-added within schools, yr 2 Average student gain per academic year No effect

Teacher 95% plausible value- added range ±.51±.71±.91 Variability in teacher value-added within schools, yr 3 Average student gain per academic year No effect

Exploring variation in trends  Which teachers and schools improved most?  Why? Under what conditions?

V. To Sum Up  The accelerated multi-cohort design is relatively easy to implement in school settings (a naturalistic data design).  It affords treatment effect results not easily obtainable through the “gold standard”— A multivariate distribution of effects linked with potential sources of their variation and dynamic over time

To Sum Up  More generally, an argument for an evolutionary, exploratory approach to accumulating evidence Data designs are now practical and analytic tools exist. Imagine if we had such information now on the 750+ schools that have been involved with LC over the past 15 years. A stronger empirical base for a design- engineering-development orientation to the improvement of schooling.

To Sum up: Useable Knowledge for Improving Schooling Anchored in: – place problems of practice improvement at the center –a working theory of practice and its improvement –Measure core work activities and outcomes –Aim for a science of performance improvement Variation is the natural state of affairs Make it an object of study Reliability is a key improvement concern in human- social resource intensive enterprise