Presentation is loading. Please wait.

Presentation is loading. Please wait.

Session 853 Developing a Framework that Integrates Program Design, Management, Accountability & Evaluation Session 853 Extending Organizational Capacity.

Similar presentations


Presentation on theme: "Session 853 Developing a Framework that Integrates Program Design, Management, Accountability & Evaluation Session 853 Extending Organizational Capacity."— Presentation transcript:

1 Session 853 Developing a Framework that Integrates Program Design, Management, Accountability & Evaluation Session 853 Extending Organizational Capacity and Capability to Evaluate Federal Environmental Research Programs Developing a Framework that Integrates Program Design, Management, Accountability & Evaluation Dale Pahl* and Emma Norland U.S. EPA Office of Research and Development October 29, 2005 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association *Corresponding author contact: pahl.dale@epa.gov

2 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Presentation Focus  Question: How can we develop a framework that  Integrates program design, management, accountability, and evaluation?  Responds to OMB guidance about the Research & Development Investment Criteria?  Communicates clearly—to evaluators, clients, and external stakeholders—about the program’s environmental research and outcomes?

3 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Presentation Focus  Response: Articulating the program theory for EPA’s environmental research creates a logical framework that...  Integrates program design, management, accountability, & evaluation;  Engages research managers, clients, scientists, and stakeholders across the program’s scope and lifetime; and  Enables independent expert panels to evaluate evidence about program relevance, quality, performance, and leadership—with client input

4 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Communicating Program Theory for Environmental Research Helps Integrate Design, Management, Accountability, & Evaluation Intermediate Outcomes Long-Term Outcomes Mission Strategic Goals & Objectives Outcomes and Environmental Results Clients Short-Term Outcomes Effective Transfer Clients Client Decisions & Actions Specific Organizations & Individuals Research Topics & Activities Research Outputs Resources Research Program Program Managers Have Direct Control Program Managers Have Direct Influence Agencies Have Indirect Impact Clients use research (short-term outcomes) … … e.g., to make environmental decisions

5 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Communicating Program Theory for Environmental Research Helps Integrate Design, Management, Accountability, & Evaluation ) Specific Clients Research Topics & Activities Key Research Outputs Resources Intended Changes in Decisions or Actions by Specific Clients Intermediate Outcomes (e.g., Improved environmental quality, reduced human exposure) Long-Term Outcomes (e.g., Improved human & ecosystem health) Short-term Outcomes Programs are designed from RIGHT to LEFT synthesis analysis

6 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Environmental Outcomes, Risk Assessment, & Accountability Adapted from Presentations to EPA’s Board of Scientific Counselors by Hugh Tilson, Larry Cupitt, and John Vandenberg  Risk assessment helps identify & prioritize scientific questions & knowledge gaps across a program’s environmental outcomes  Risk assessment is essential to help:  Decide whether or not to take regulatory action  Decide whether or not to take regulatory action... Is there an environmental hazard?  Decide what actions are most effective  Decide what actions are most effective... What actions do we take to protect human health?  Understand how to evaluate the effectiveness of our decisions  Understand how to evaluate the effectiveness of our decisions... Were we effective? Ambient Conditions Fate and Transport Source Emissions Exposure And Dose Early Signs of Effects Health Impacts

7 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Environmental Outcomes, Risk Assessment, & Accountability Environmental Outcomes, Risk Assessment, & Accountability Adapted from Risk Assessment in the Federal Government: Managing the Process (NRC, 1983); 1997 Update to ORD’s Strategic Plan (EPA, 1997); and OIG-ORD Presentation to EPA’s Deputy Administrator (Pahl & Norland, March 2002) EXPOSURE RISK ASSESSMENT RISK MANAGEMENT 3. Accountability Developing and measuring appropriate environmental indicators demonstrates whether environmental decisions result in improved human and environmental health. Pollution Sources Emissions Transport & Transformation Environmental Concentrations ExposuresExposure Assessment Risk Characterization Legal Considerations Ecosystem and Human Health Considerations Social, Economic, & Political Factors Exposure-Dose Relationships Dose- Response Assessment Risk Management Options 2. Implementation Decisions Managers make decisions about how to implement, comply with & enforce regulations or how to remedy environmental problems. Internal Dose Biological Effect Adverse Health Effect Health Assessment Hazard Identification Dale Pahl Nov ‘02 1.Environmental Decisions and Regulations

8 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Short-Term Outcomes: A Critical Link between Research & Impact Intended Changes in Decisions or Actions by Specific Clients Limitations or Gaps in Knowledge about the Environmental Problem Limitations or Gaps in Attitudes Limitations or Gaps in Skills and Abilities Needed to Respond to the Environmental Problem Transfer (e.g., dissemination thru publications) Transfer (e.g., guideline manuals & training) Specific Clients Short-Term Outcomes (Research Contributions to Environmental Decisions) Intermediate Outcomes (e.g., improved environmental quality, reduced human exposure) Long-Term Outcomes (e.g., improved human & ecosystem health) Research Topics & Activities Research Outputs

9 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Short-Term Outcomes: A Critical Link between Research & Impact   Short-term outcomes are achieved when key research contributions are transferred to, and used by, intended clients   Research creates improved knowledge and applications for risk assessment, environmental decisions, regulatory decisions, and accountability: Setting or revising standards—formal EPA rulemaking Implementing standards—e.g., Regions or states develop plans to comply with standards, restore ecosystems, or manage environmental exposure & risk Applying environmental indicators to “measure” progress to achieve regulations and environmental outcomes—e.g., assessing whether legislation & regulations have the intended environmental impact ____ ____ 1 See, for example: Strategic Research Plan for Particulate Matter, Air Quality Subcommittee of the Committee on the Environment and Natural resources (CENR), December, 2002; Science to Support Rulemaking, EPA Office of Inspector General (OIG) Report 2003-P-00003, November 15, 2002; and Air Quality Management in the United States, National Research Council, 2004.

10 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Communicating Program Theory for Environmental Research Helps Integrate Design, Management, Accountability, & Evaluation ) Specific Clients Research Topics & Activities Key Research Outputs Resources Intended Changes in Decisions or Actions by Specific Clients Intermediate Outcomes (e.g., improved environmental quality, reduced human exposure) Long-Term Outcomes (e.g., improved human & ecosystem health) Short-term Outcomes Programs are designed from RIGHT to LEFT synthesis analysis

11 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Organizing Research, Topics, & Activities   Both applied 1 and use-inspired basic research 2 contribute to regulatory decisions and improved understanding about environmental outcomes: Applied 1 environmental research is targeted at understanding and solving particular environmental problems. Basic 2 environmental research elucidates problems that involve complex environmental processes and nonlinear systems with multiple causes and effects. EPA’s basic research is “use-inspired” basic research 2 —targeted to answering questions about complex problems that cut across EPA’s programs to ensure that decisions are based on a foundation of sound science.   Typically, the distinction between these types of research is not clear cut. __________ __________ 1 EPA's applied and basic research programs implement recommendations from the National Research Council in Building a Foundation for Sound Environmental Decisions, (NRC, 1997). 2 Stokes, D.E. “Renewing the Compact between Science and Government,” in 1995 Forum Proceedings, Vannevar Bush II—Science for the 21 st Century. Pages 15-32. Sigma Xi, 1995.

12 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Organizing Research, Topics, & Activities Adapted from National Research Council, Building a Foundation for Sound Environmental Decisions (NRC, 1997). Use risk assessment to rank issues and pinpoint largest uncertainties Identify existing and emerging issues for a specific problem Narrow EPA focus based on client needs and recognition of what others are doing Identify research topics that improve understanding, reduce uncertainties, and develop client applications Problem-Driven Research Core Research Select projects based on broad applicability, relevance to EPA, and scientific merit. Elucidation of Complex Environmental Processes Development of Tools Collection of Data FEEDBACK Re-evaluate priorities regularly

13 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Organizing Research, Topics, & Activities   Question: How do we organize and “measure” research when there are no objective methods to:  Measure new knowledge as it develops?  Manage the pace at which research progresses?  Measure research quality and impact?

14 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Organizing Research, Topics, & Activities   Response: Organize the research with priority research topics Assess research progress and priorities with periodic meetings Convene independent expert panels to evaluate the improved knowledge and its applications with indicators that span a program’s scope & lifetime __________ 1 For example, see Averch, H.A., “The Systematic Use of Expert Judgment,” pages 294-295 in Handbook of Practical Program Evaluation. J.S. Wholey, H.P. Hatry, and K.E. Newcomer (eds). San Francisco, Jossey-Bass, 1994. 2 The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven. U.S. GAO/GGD (1997)

15 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 Developing Indicators Across a Program’s Scope & Lifetime Potential Indicators: Evaluation Criteria Quality Relevance Performance Scientific Leadership Peer Review (multi-year research plan) Peer Review (topic / project / activity) Bibliometric Analysis (publications) Client Feedback (client use of research) Independent Expert Review (including client feedback) Very useful UsefulLimited usefulness

16 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 The Importance of Independent Expert Evaluation   The systematic use of independent expert judgment is important 1,2 when:  It is difficult to measure program progress or outputs (e.g., advances in research knowledge) needed to achieve outcomes  Multidisciplinary expertise is needed to evaluate scientific progress that responds to research topics and scientific questions  It is difficult to determine when outcomes can be attributed to the program  Agency programs involve research, regulation, or external partners such as state agencies  These criteria illustrate why independent expert review is important for evaluating EPA’s research programs _____ 1 For example, see Averch, H.A., “The Systematic Use of Expert Judgment,” pages 294-295 in Handbook of Practical Program Evaluation. J.S. Wholey, H.P. Hatry, and K.E. Newcomer (eds). San Francisco, Jossey-Bass, 1994. 2 The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven. U.S. GAO/GGD (1997)

17 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05Summary:  EPA follows a systematic approach 1,2 to organize, integrate, synthesize, and evaluate research that informs environmental decisions by specific clients: Develop an integrated risk assessment framework to synthesize available information about a specific environmental problem Assess client needs, knowledge gaps, research questions, and uncertainties Create research topics needed to develop knowledge & evaluate evidence related to the gaps, questions, and uncertainties Monitor progress in implementing the research topics and in applying improved knowledge base Evaluate the improved knowledge base and client applications  The next slide illustrates how program theory creates a framework that integrates these objectives... __________ 1 For an example of EPA’s systematic approach, see Strategic Research Plan for Particulate Matter, Air Quality Subcommittee of the Committee on the Environment and Natural resources (CENR), December, 2002. 2 For additional background information,see Risk Assessment in the Federal Government: Managing the Process. National Research Council, 1983; and Environmental Research and Development: Strengthening the Federal Infrastructure. Carnegie Commission on Science, Technology and Government, 1992.

18 2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association Pahl & Norland Oct’05 A risk-based assessment of knowledge gaps & needs related to EPA’s strategic goals— including strategic guidance from independent advisors such as the NAS The number, sequence, and distribution of research projects / activities that respond to the priority research topics Peer-reviewed publications that respond to the priority research topics, as measured by bibliometric analysis The program integrates new knowledge and key research contributions for application at key decision-points (e.g., regulations) by specific clients Client use of research knowledge and applications at key decision-points The information-value or decision-value of the research knowledge or applications for understanding complex environmental processes The information-value or decision-value of the research knowledge for developing indicators of public health or ecosystem health Research Coordination Teams assess and prioritize client needs for research to inform policy, decision-making, and accountability The number of research projects / activities completed during the past 5 years that respond to the priority topics The information value of the publications (for example, in developing new knowledge or in reducing uncertainty) as measured by bibliometric analysis The program collaborates with clients to transfer & demonstrate key research contributions The information-value or decision-value of the research knowledge and applications used by clients The information-value or decision-value of the research knowledge or applications for measuring environmental progress, developing risk mitigation approaches, or evaluating the effectiveness of regulations The relevance of the program’s outcomes to EPA’s mission and legislative mandates Research Coordination Teams select priority topics or questions to organize the research that leads to outcomes The program ensures high quality research with merit-based competitive awards (e.g., through grant mechanisms) that respond to the priority research topics Research Coordination Teams assess client feedback on key research contributions Progress to achieve outcomes is measured with a small number of long-term goals and measures The information-value or decision-value of the research knowledge or applications for helping EPA to achieve its strategic goals. Partnerships provide FTE or $ to support priority research topics linked to EPA outcomes & strategic goals The program sponsors periodic meetings to assess ongoing research progress and priorities Scientific leadership combines several elements: (1) The program has developed priority research topics that help EPA achieve its strategic goals; (2) The program has developed partnerships that support & coordinate research which focuses on one or more of these topics; (3) Bibliometric analysis indicates that the program has made significant contributions that respond to the priority research topics; and (4) The program’s principal investigators are recognized leaders in their research disciplines Evidence about program relevance, quality, performance, and leadership—evaluated by independent expert panels Long-Term Outcomes Environmental Outcomes and Strategic Goals Resources & Inputs Research Topics & Activities Research Outputs Specific Clients Short-Term Outcomes WHAT? The right science HOW? Programs are designed from RIGHT TO LEFT WHY? The right questions


Download ppt "Session 853 Developing a Framework that Integrates Program Design, Management, Accountability & Evaluation Session 853 Extending Organizational Capacity."

Similar presentations


Ads by Google