< This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of.

Slides:



Advertisements
Similar presentations
Alternative Approaches to Campsite & Trail Monitoring Jeff Marion, Unit Leader/Scientist Virginia Tech Field Unit, USGS, Patuxent WRC
Advertisements

Intelligence Step 5 - Capacity Analysis Capacity Analysis Without capacity, the most innovative and brilliant interventions will not be implemented, wont.
Technical Committee Training Prof Richard Mackay, AM June 2012 & Management of Tourism Carrying Capacity.
 Is extremely important  Need to use specific methods to identify and define target behavior  Also need to identify relevant factors that may inform.
PROJECT RISK MANAGEMENT
Mywish K. Maredia Michigan State University
As presented to the Global Colloquium on Engineering Education Deborah Wolfe, P.Eng. October 2008 The Canadian Process for Incorporating Outcomes Assessment.
Chapter 15 Evaluation Recognizing Success. Social Work Evaluation and Research Historically –Paramount to the work of early social work pioneers Currently.
Scoping in the EIA process Proposal Identification Screening
Project Monitoring Evaluation and Assessment
2011 OSEP Leadership Mega Conference Collaboration to Achieve Success from Cradle to Career 2.0 State Monitoring Under IDEA A Snapshot of Past Practices.
Statistical Issues in Research Planning and Evaluation
Mitigating Risk of Out-of-Specification Results During Stability Testing of Biopharmaceutical Products Jeff Gardner Principal Consultant 36 th Annual Midwest.
This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of interest,
Introduction to Research
Specifying a Purpose, Research Questions or Hypothesis
< This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of.
Chapter 3 Preparing and Evaluating a Research Plan Gay and Airasian
Chapter Three Research Design.
PPA 503 – The Public Policy Making Process
Developing the Marketing Plan
Capability Maturity Model
How to Develop a Project Evaluation Plan Pat Gonzalez Office of Special Education Programs
 This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of.
Literature Review and Parts of Proposal
N By: Md Rezaul Huda Reza n
Vision, Goals, Objectives, and Alternatives. Objectives  Define vision, goals, objectives, and alternatives within the planning context.  Describe collaborator.
Indicators Dr Murali Krishna Public Health Foundation of India.
1 What are Monitoring and Evaluation? How do we think about M&E in the context of the LAM Project?
Chapter 6 : Software Metrics
Writing research proposal/synopsis
< This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of.
Applying Social Science to Outdoor Recreation Management Diane Kuehn SUNY ESF.
Carrying Capacity (CC) and LAC
Understanding customer expectations and perceptions
© 2011 Underwriters Laboratories Inc. All rights reserved. This document may not be reproduced or distributed without authorization. ASSET Safety Management.
Research Design.
Management & Development of Complex Projects Course Code MS Project Management Perform Qualitative Risk Analysis Lecture # 25.
 2008 Johns Hopkins Bloomberg School of Public Health Evaluating Mass Media Anti-Smoking Campaigns Marc Boulay, PhD Center for Communication Programs.
This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of interest,
Division Of Early Warning And Assessment MODULE 5: PEER REVIEW.
ASSESSING AND MANAGING WILDLAND RECREATIONAL DISTURBANCE Stephen J. DeMaso, Fidel Hernández, and Leonard A. Brennan Caesar Kleberg Wildlife Research Institute,
PREMISES FOR DEVELOPING AND APPLYING SEDIMENT QUALITY OBJECTIVES Presented and (mostly) agreed upon during the October 27, 2004 meeting of the Advisory.
Integrated Risk Management Charles Yoe, PhD Institute for Water Resources 2009.
Quantitative and Qualitative Approaches
STRATEGIC ENVIRONMENTAL ASSESSMENT METHODOLOGY AND TECHNIQUES.
Screen 1 of 22 Food Security Policies – Formulation and Implementation Policy Monitoring and Evaluation LEARNING OBJECTIVES Define the purpose of a monitoring.
Evidence-based Education and the Culture of Special Education Chair: Jack States, Wing Institute Discussant: Teri Palmer, University of Oregon.
This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of interest,
Begin at the Beginning introduction to evaluation Begin at the Beginning introduction to evaluation.
1 of 36 The EPA 7-Step DQO Process Step 6 - Specify Error Tolerances (60 minutes) (15 minute Morning Break) Presenter: Sebastian Tindall DQO Training Course.
SOFTWARE PROJECT MANAGEMENT
An Expanded Model of Evidence-based Practice in Special Education Randy Keyworth Jack States Ronnie Detrich Wing Institute.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
3-1 Copyright © 2010 Pearson Education, Inc. Chapter Three Research Design.
Research Methods Observations Interviews Case Studies Surveys Quasi Experiments.
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
1 of 31 The EPA 7-Step DQO Process Step 6 - Specify Error Tolerances 60 minutes (15 minute Morning Break) Presenter: Sebastian Tindall DQO Training Course.
UNIT 7 MONITORING AND EVALUATION  Monitoring and evaluation is the process of examining progress against institution’s goals or plan.  The term SM &
Basic Concepts of Outcome-Informed Practice (OIP).
Copyright © 2011 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 1 Research: An Overview.
1 Formulate Alternatives Planning Step 5. 2 Social Science Activities in Land Use Planning Planning Steps Social Science Activities Steps 1 & 2: Identify.
1 of 48 The EPA 7-Step DQO Process Step 6 - Specify Error Tolerances 3:00 PM - 3:30 PM (30 minutes) Presenter: Sebastian Tindall Day 2 DQO Training Course.
Organizations of all types and sizes face a range of risks that can affect the achievement of their objectives. Organization's activities Strategic initiatives.
Research Philosophies, Approaches and Strategies Levent Altinay.
EIAScreening6(Gajaseni, 2007)1 II. Scoping. EIAScreening6(Gajaseni, 2007)2 Scoping Definition: is a process of interaction between the interested public,
Chapter Three Research Design.
What Is Planning? According to Koontz & O’Donell,
MONITORING AND EVALUATION IN TB/HIV PROGRAMS
Standards-based Individualized Education Program (IEP) Module Two: Developing the Present Level of Academic Achievement and Functional Performance (PLOP)
Presentation transcript:

< This document is contained within the Visitor Use Management Toolbox on Wilderness.net. Since other related resources found in this toolbox may be of interest, you can visit this toolbox by visiting the following URL: c=vum. All toolboxes are products of the Arthur Carhart National Wilderness Training Center. c=vum c=vum

Identifying Indicators & Standards Jeff Marion, Unit Leader/Scientist Virginia Tech Field Unit, USGS, Patuxent WRC

Selecting Indicators & Standards Presentation Objectives: 1. Define indicators and describe their role, review criteria for their selection, and illustrate with trail impact examples. 2. Define indicator standards and describe their role, describe alternative types of standards, and illustrate with trail and campsite examples.

LAC & VERP Planning And Management Decision Making Frameworks Standards Exceeded Select Appropriate Management Action Establish Prescriptive Management Objectives Choose Biophysical and Social Indicators of Change Formulate Standards Monitor Conditions Compare Conditions to Standards Standards Not Exceeded Evaluate and Identify Causal Factors Implement Management Action

What Are Indicators? What are indicators and why do we use them?  Indicators are measurable physical, ecological, or social variables used to track trends in conditions caused by human activity so that progress toward goals / desired conditions can be assessed.  Indicators provide a means for restricting information collection and analysis to the most essential elements needed to answer management questions.

Indicator Selection: A Process Begin by defining monitoring questions: Begin by defining monitoring questions:  “Are visitors experiencing an environment where the evidence of human activity is substantially unnoticeable?”  “Are trails in good usable condition?” or, “Are the number and length of visitor-created trails increasing?”

Indicator Selection: Sources of Indicators Employ one or more of the following sources:  Consult Scientific Literature: There is a large and growing body of applicable literature.  Conduct Scientific Research: Conduct site-specific research or inventories where necessary to address information gaps.  Consult the Public: Involve stakeholders.  Apply Management Judgment: Consult with experienced managers, both local and distant.

Indicator Selection: Screening Questions  Does the indicator tell you what you want to know? Is it ecologically or experientially meaningful?  Is the indicator independent of environmental variation?  Can the indicator be measured reliably?  Is the indicator responsive to manage - ment action within a reasonable time frame?

Indicator Selection: Screening Questions  Does the indicator act as an early warning, alerting you to deteriorating conditions before unacceptable change occurs?  Can the indicator be measured with minimal impact to the resource or the visitor’s experience?  Can the indicator provide information worth the time and cost required?

Indicator Selection: Screening Criteria Primary Criteria Secondary Criteria Specific Easy to train for monitoring ObjectiveCost-effective Reliable and repeatable Minimal variability Related to visitor use Responds over a range of conditions Sensitive Large sampling window Resilient Availability of baseline data Low-impact Significant (Source: VERP Handbook, 1997)

Trail Indicators  Selected from a review of the scientific and management literature.  LAC/VERP frameworks typically include a small number of indicators, far fewer than those shown in the following lists.  Indicator assessment/monitoring methods need to be considered during the indicator selection process.

Forms of Trail Impacts  Informal trails (number, length, condition)  Trail widening  Soil erosion  Trail muddiness

Trail Indicators Monitoring approaches must be considered in selecting indicators and setting standards. Two common types:  Problem Assessment: Assess location and lineal extent of all occurrences of pre-defined tread impacts (e.g., eroded or muddy sections).  Point Sampling: Assess tread conditions through individual measures of indicators at sampling points.

Trail Assessment Survey Types Point Sampling Problem Assessment

PA - Problem Assessment, PS - Point Sampling Common Trail Indicators Soil Erosion  Cross Sectional Area (sq ft) PS  Maximum Incision (in) PS  Excessive Erosion (# ft/mi, lineal ft, % of trail length) PA T2T2 T3T3 T5T5 I2I2 I3I3 I4I4 I1I1 I6I6 T4T4 I5I5 T1T1 T6T6

PA - Problem Assessment, PS - Point Sampling Common Trail Indicators Tread Muddiness  Muddiness (% of tread width) PS  Excessive Muddiness (# ft/mi, lineal ft, % of trail length) PA Tread Width  Tread Width (ft) PS  Excessive Width (# ft/mi, lineal ft, % of trail length) PA

PA - Problem Assessment, PS - Point Sampling Common Trail Indicators Multiple Treads  Multiple Treads (# at sample points) PS  Multiple Treads (# ft/mi, lineal ft, %) PA Visitor-Created Trails  Informal Trails (#, #/mi, lineal extent, % of formal trail miles)  Other concerns related to informal trails: density, proximity to sensitive resources, and condition

What Are Indicator Standards? What are indicator standards and why do we use them? Indicator Standards - Measurable statements that define minimally acceptable conditions. Standards narrow the focus of management from broad, qualitative statements of management intent to specific statements of desired conditions that provide a basis for decision making.

What Are Indicator Standards? Standards reflect desired conditions. Standards:  Are specific and achievable.  Describe ends (desired resource and social conditions) rather than means (management actions). Standards are established for setting attributes that reflect degree of naturalness or that influence experience quality.  Setting attributes must be subject to management control. For example, a standard could be written for encounter rates but not for solitude or satisfaction, which are direct attributes of the experience.

What Are Indicator Standards? Standards quantitatively specify a limit on the amount of change that will be tolerated.  Standards define the compromise between resource/visitor experience protection and access to recreational opportunities, i.e., they represent an explicit trade-off between two conflicting goals.  Standards are statements of minimally acceptable conditions (thresholds of acceptability). They define the point past which conditions become unacceptable, they do not define desired or unacceptable conditions.

Characteristics of Good Standards  Quantitative: Measurable.  Time or Space-Bounded: Expressed as “x”/day or “y”/unit area.  Expressed as a probability: Allows flexibility when needed for random or peak-use events.  Impact-Oriented: Focus on impacts, not on management actions (e.g., conditions, not # of visitors)  Realistic: Focus on attainable conditions.

Interpreting Standards Standards are absolute limits, not just warnings.  Violation of standards should not be tolerated.  Tolerances can be written into standards in the form of probabilities. e.g., <10 encounters/day on 90% of the summer use season days. Such standards allow for more crowded conditions during a few peak use weekends.  When a standard is exceeded managers should implement actions to avoid compromising resource or experience conditions further.

Interpreting Standards Standards are value-based management decisions.  Setting standards ultimately involves inherently subjective and value-based management decisions.  It is desirable that standards be developed through a collaborative process with input from stakeholders.  Public involvement and/or participation is preferred but managers have been delegated the authority and responsibility for making final decisions.

Setting Standards: Role of Science Science can characterize the range of indicator conditions to allow evaluation of draft standards. Indicator data may also be examined to look for “thresholds”:  e.g., level of vegetation disturbance beyond which annual recovery cannot occur. Standards can be informed by science but cannot be derived from science.  Standards are evaluative, they must consider resource impacts in the context of the recreational “costs” of not exceeding them.

Types of Standards Maximum Condition - Specifies maximum acceptable indicator value at a single site. e.g., trail width < 10 in Aggregate Condition - Specifies maximum acceptable aggregate indicator value for a group of sites. e.g., Sum of informal trail lengths in travel zone x < 3.5 mi Rate of Change - Specifies maximum acceptable rate of change between monitoring cycles. e.g., increase in trail incision < 0.25 in/yr

Examples of Standards: Trails Potential IndicatorsExamples of Standards Informal trails/unit area <2500 ft/5 acres Increase in informal trail lengths < 100% Trail width<4 ft Trail width >3 ft 3 ft<500 ft/mi Soil erosion <2 ft Soil erosion > 2 ft 2 ft<300 ft/mi Trail muddiness<5% of trail length Note: Standards are generally set and vary by zones.

Examples of Standards: Campsites Potential IndicatorsExamples of Standards Campsites/unit area <12 sites/5 acres Campsite size/site<1800 ft 2 Campsite size/unit area<8000 ft 2 /5 acres Increase in campsite size < 100% Soil exposure/site<1200 ft 2 /campsite Soil exposure/unit area <6000 ft 2 /5 acres Tree damage <10 damaged trees/site Fire sites<3/site Note: Standards are generally set and vary by zones.

Conclusions  Scientific literature and research can help identify alternative resource indicators and provide objective measurement methods  Scientific surveys can describe the distribution of values for each indicator – illustrating outcomes of alternative standards.  Science cannot be used to identify indicator standards. Standards are evaluative, they represent tradeoffs between competing management objectives.