Data for PRS Monitoring: Institutional and Technical Challenges
Institutional Preconditions PRS strategy is in place and objectives have been clearly set Monitoring effort wasted – garbage in, garbage out Evaluation can reveal wasted effort Agreement on where monitoring should occur are policies in place? are inputs being provided? are outputs being achieved? are outcomes being realized? Define specifics of what should be monitored and what should be evaluated
Why “M” and “E”? Management perspective Project/program perspective Oversight and Accountability Decisions about future funding based on the results Desired effects Implementation Assure public and stakeholders that money is spent appropriately Ensure staff or contractor accountability Learning Enhance learning among stakeholders Project/program perspective Design and Implementation Feed back of results to improve implementation
Some Definitions: “M” A continuing function that uses systematic collection of data on specified indicators Provides management with indications of the extent of progress toward milestones and targets (results - MDGs) achievement of outputs and outcomes use of funds Generally comprehensive for, and conducted by, business unit
Some Definitions: “E” A periodic function that uses systematic assessment of intended and unintended consequences of a specific intervention Provides management and the main stakeholders with indications of quality, effectiveness, efficiency attribution of effects to the intervention key information for decision-making Conducted by independent evaluators
Defining the Scope Purpose -- how to use M&E data? Accountability? Planning? Learning? Better Implementation? Budgeting? Prioritization – what data are important to collect? Organization – who will collect data? Periodicity – when? Reporting – how the data be turned into usable information? Costs – what are the cost implications for data collection and use? Relevance, reliability, and validity of data– how will it be ensured? Coordination of data– what are the mechanisms for interagency collaboration?
Being “SMART” About Monitoring Indicators Specific Measurable Achievable Relevant Timely Valid – measure what they set out to measure Reliable – different people/agencies must produce the same statistic Cost-Effective Indicator have some key characteristics. First, they are SMART.