Risk Evaluation: Maximizing Risk Accuracy Presentation to Special Commission to Reduce the Recidivism of Sex Offenders 10/8/2014.

Slides:



Advertisements
Similar presentations
Conceptual Issues in Risk Assessment Randy K. Otto, PhD Department of Mental Health Law & Policy Florida Mental Health Institute University of South Florida.
Advertisements

Standardized Scales.
ASSESSING RESPONSIVENESS OF HEALTH MEASUREMENTS. Link validity & reliability testing to purpose of the measure Some examples: In a diagnostic instrument,
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Risk Assessment in the SVP Context Natalie Novick Brown, PhD, SOTP th St. NE, Suite 201 Seattle, Washington
Chapter 10 Decision Making © 2013 by Nelson Education.
California Static Risk Assessment (CSRA)
CSOM Training Curriculum: An Overview of Sex Offender Treatment for a Non-Clinical AudienceLong Version: Section 21 Sex Offender-Specific Treatment Outcome.
Sex Offender Treatment US Probation Central California Presented by Helene Creager, LCSW Supervisor & Mental Health Coordinator US Probation Central District.
Sex offenders: Treatment & risk assessment
Risk Evaluation: Maximizing Risk Accuracy MATSA/MASOC Presentation to SORB 1/31/2013.
Chapter 4 Validity.
Copyright © 2012 Pearson Canada Inc Chapter 10 Risk Assessment.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Jan Weiss, PT, DHS, CLT-LANA
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Sex Offenders. Sex Offenders… Contact Offenders – male victims Contact Offenders – female victims Non-contact Offenders – paraphilia Rapists Child molesters.
Classroom Assessment A Practical Guide for Educators by Craig A
Sexual Offenders: What the Research Reveals
Assessment of Sex Offenders. Learning Objectives Identify information and assessments that reliably estimate risk posed by sex offenders; Describe some.
TAYLOR HOWARD The Employment Interview: A Review of Current Studies and Directions for Future Research.
Copyright © 2008 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. John W. Creswell Educational Research: Planning,
JUVENILE SEX OFFENDER ASSESSMENT PROTOCOL J-SOAP II WJCIA ANNUAL CONFERENCE THURSDAY, SEPT STEVENS POINT, WISCONSIN.
Joe Judge.  There are significant literatures on risk factors for recidivism in sexual offenders and on the predictive accuracy of different types of.
Psychopathy, Violence Risk Assessment, and the Personality Assessment Inventory (PAI) Mark Hastings, Jeff Stuewig, Amy Drapalski, & June Tangney George.
Criminal Psychology Chapter 6 From Dangerousness to Risk Assessment Talbot Kellogg Community College.
American Psychology-Law Society March 2013 Portland, OR 1 Partially Specified Actuarial Tables and the Poor Performance of Static-99R Richard Wollert Ph.D.
Determining Sample Size
Evidence-Based Sentencing. Learning Objectives Describe the three principles of evidence- based practice and the key elements of evidence-based sentencing;
Raymond Knight, Ph.D. 1 Civil Commitment: Dubious Solution to a Serious Problem?
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
CSOM Training Curriculum: An Overview of Sex Offender Treatment for a Non-Clinical AudienceShort Version: Section 21 Describe the general findings of sex.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Evaluating a Research Report
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
Reliability & Validity
2010 Annual State of Hawaii Forensic Mental Health Examiner Training Conference, Kaneohe, Hawaii Quality of Conditional Release Reports Submitted to the.
Salient Factor Score CTSFS99. What it is How to use it.
Classroom Assessment A Practical Guide for Educators by Craig A. Mertler Chapter 13 Assessing Affective Characteristics.
Community Notification, Risk Assessment, and Civil Commitment of Sex Offenders.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Psychopathy and Criminal Recidivism in Female Offenders A 10-Year Follow-up of a Nationwide Sample Weizmann-Henelius, Ghitta Grönroos, Matti Eronen, Markku.
Discussion of Risk Assessment Presentation Raymond Knight.
Educable Mental Retardation as a Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
QUANTITATIVE RESEARCH Presented by SANIA IQBAL M.Ed Course Instructor SIR RASOOL BUKSH RAISANI.
Assessment Tools and Community Supervision of Sexual Offenders Robin J. Wilson, PhD, ABPP Chris Thomson, M.A.
Offender Risk Management Model Community Forum Teal Maedel Psychologist Vancouver Parole RCMP Behavioural Sciences Group.
Tier III Implementation. Define the Problem  In general - Identify initial concern General description of problem Prioritize and select target behavior.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Chapter 13 Repeated-Measures and Two-Factor Analysis of Variance
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Assessing Responsiveness of Health Measurements Ian McDowell, INTA, Santiago, March 20, 2001.
How do we know whether criminals will re-offend?.
Proposed SORB Regulations Risk Factors Critique Raymond Knight, Ph.D.
Chapter 6 - Standardized Measurement and Assessment
Reliability a measure is reliable if it gives the same information every time it is used. reliability is assessed by a number – typically a correlation.
What We Know About Assessment of Risk of Recidivism and Criminogenic Needs of Offenders: Why and How to Do Assessments? Robin J. Wilson, PhD, ABPP
Thinking About A Risk-Based Registry. Sex offender risk assessments are most often employed in applied forensic settings for purposes of decision-making.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Sex offender risk assessments in the child protection context: Helpful or not? Ms Karen Broadley Child Abuse Prevention Research Australia.
Data Collection Methods NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
Sexual Offenders Chapter 6.
Concept of Test Validity
Method Results Discussion
Presentation transcript:

Risk Evaluation: Maximizing Risk Accuracy Presentation to Special Commission to Reduce the Recidivism of Sex Offenders 10/8/2014

Overview of Presentation Brief history of risk assessment and the different kinds of assessment that have been developed; Indication of where MA SORB Classification fits in this historical context, and in the context of current state strategies; Summary of the criteria for how one should evaluate risk instruments; Quick overview of the recent empirical evaluations of risk instruments; Suggestions of two strategies for improving classification in MA.

BRIEF HISTORY OF RISK ASSESSMENT

Brief History First generationFirst generation – Unstructured clinical judgment, including structured clinical guidelines (SCG). Second generationSecond generation – Actuarial risk scales comprising static, historical factors. Third generationThird generation – the assessment of “criminogenic needs” or dynamic risk factors. Bonta, 1996 Fixed or historical factors that cannot be changed (such as age at first offense) Potentially changeable factors, both stable, but potentially changeable risk traits, and acute, rapidly changing factors.

Brief History Characteristics of Unstructured Clinical Judgments –Characteristics of Unstructured Clinical Judgments – No items specified for considering risk level; Method for combining items is not specified. (Hanson & Morton-Bourgon, 2009) First Generation

Brief History Characteristics of SCGs–Characteristics of SCGs– They identify items to use in the decision and typically provide numerical values for each item; Although they also usually provide a method for combining the items into a total score, they do not specify a priori how the clinician should integrate the items; No tables linking the summary scores to recidivism rates. (Hanson & Morton-Bourgon, 2009) First Generation

Brief History Requirements of Empirical Actuarials –Requirements of Empirical Actuarials – Provide specific items to make the decision with quantitative anchors, which are derived from empirical investigation; Method for combining the items into an overall score is specified; Tables linking the summary scores to recidivism rates are provided. (Hanson & Morton-Bourgon, 2009) Second Generation

Brief History Requirements of Mechanical Actuarials – They provide specific items for the decision with numeric values for each item, which are derived from a review of literature and theory; Method for combining the items into an overall score is specified; Tables linking the summary scores to recidivism rates are n nn not provided. (Hanson & Morton-Bourgon, 2009) Second Generation

Brief History Additional condition Adjusted Actuarials –Additional condition Adjusted Actuarials – Use appropriate actuarials (empirical or mechanical); The clinician adjusts the score (and the recommendation) using factors external to the actuarial. (Hanson & Morton-Bourgon, 2009) Second Generation

MA SORB CLASSIFICATION FACTORS Where Does It Fit?

MA SORB Classification Factors Somewhere between an unstructured judgment and an SCG –Somewhere between an unstructured judgment and an SCG – Where Does It Fit? Predictive Validity AWAcrimeAWAcrimeClinJudgmtClinJudgmtSCGSCGEmp.ActuarialEmp.Actuarial Em. Act. + Dyn. Em. Act. + Dyn. MASORBMASORB

MA SORB Classification Factors Somewhere between an unstructured judgment and an SCG –Somewhere between an unstructured judgment and an SCG – butIt specifies a set of factors to be considered; but does not provideIt does not provide any quantification of these factors (i.e., numeric item scores). notIn many items it does not provide clear specification of where the cutoff for “presence” or “absence” of a factor would be. Thus, it provides limited guidance both on the presence of items and on the combining of items. Why Does It Fit Here?

 Item 3. Psychopathy Code this by reference to the PCL-R. Code PCL- R scores of 30 or above as “Y,” scores of as “?,” and scores of 20 or lower as “N.” Y = 2 ? = 1 N = 0 Example of SCG MA SORB Classification Factors SVR-20

 Item 2. Repetitive and Compulsive Behavior Example of SORB Factors ?charges, convictions, self-report? ?includes both impulsive and compulsive behavior? MA SORB Classification Factors Could be either NoScore VagueCriteria&NoCutoff

So the MA SORB criteria neither— provide a metricprovide a metric for each item, so it is not known which items an expert is depending on and no item improvement can be attempted, nor specify the cutoffspecify the cutoff criteria necessary for items to be judged present or absent by two raters, so no determination of agreement or reliability can be ascertained. rules on how to combineMoreov er, there are no rules on how to combine or weight items in reaching a decision. MA SORB Classification Factors

Relative to other states?Relative to other states? Where Does It Fit?

Identified “Tiering”

De Facto “Tiering”

Criteria for De Facto “Tiering” 6% State Actuarial

Criteria for De Facto “Tiering” 6% State Actuarial

MN Leveling Criteria Actuarial Leveling Criteria Clinical Judgment Trumps 6% Hx of gratuitous violence Unsuccessful treatment Predatory offense behavior Supervision failures

HOW DO WE EVALUATE RISK TOOLS? Evaluating Reliability and Validity

Reliability HOW DO WE EVALUATE RISK TOOLS?

Reliability Accuracy Freedom from variable error Consistency Across raters Across items Across different measures of the same construct Across time Reliability is --

Reliability Interrater

Interrater Reliability R 1 R 2 Agreement High Reliability Low Reliability Disagreement

Reliability Interrater Internal Consistency

Agreement or Correlation Among Items = High Reliability

Allows one to calculate various forms of reliability – Item reliability Reliability of subscales (e.g., sexual deviance, criminality, etc.) Internal consistency of items in the instrument Thus, quantification allows us to restructure items and their anchors to improve reliability. Advantages of Quantification Allows Reliability Checks Gives us the Power of Being on the Same Page

Most popular SCGs and actuarials assessed in the comparative literature have acceptable reliability. Unstructured judgments have poor reliability. The reliability of MA SORB Classification Factors have not and can not be assessed. SCGs and Actuarials Reliability Results

HOW DO WE EVALUATE RISK TOOLS? Evaluating Reliability and Validity

Validity HOW DO WE EVALUATE RISK TOOLS? Validity

Validity Answers the Question Does a test measure what it is suppose to measure? What does a test measure? What can one do with the test? What does a test score predict?

Validity Answers the Question Does a test measure what it is suppose to measure? What does a test measure? What can one do with the test? What does a test score predict?

Predicting Sexual Recidivism Instrument Typed(95% CI) Empirical Actuarial.67( ) Mechanical Actuarial.66( ) SCG.46( ) Unstructured Judgmt.42( ) (Hanson & Morton-Bourgon, 2009)

significantly betterOverall, controlling for a large number of study variables, Empirical and Mechanical were significantly better predictors of recidivism; SCGs using clinical judgment and SCGs that calculate total scores did not differ. loweredIn all studies examined, clinicians’ adjustment of actuarial scores consistently lowered predictive accuracy. Predicting Sexual Recidivism (Hanson & Morton-Bourgon, 2009)

Across multiple areas of prediction, mechanical actuarial prediction (statistical prediction rules [SPRs]) has been shown to be superior to clinical judgment. A recent meta-analysis summarizes the results of years of research (Grove et al., 2000). Why Is Clinical Judgment Inferior?

All studies published in English from 1920s to mid 1990s. 136 studies on the prediction of health- related phenomena or human behavior. (Grove et al., 2000)

A large body of research has documented the reasons for the cognitive errors that clinicians make. For instance, clinicians are great at making observations and rating items, but they, like all humans, are worse than a formula at adding the items together and combining them. Why Is Clinical Judgment Inferior?

Allows one to use various strategies for improving validity of a measure– Assess item correlation with outcome; Adjust item cutoffs to maximize prediction; Assess the validity of subscales (e.g., sexual deviance, criminality, etc.); Optimize item weights for decision-making and predicting. Thus, one can restructure items, their anchors, cutoffs, and combinations to improve validity. Advantages of Quantification Allows Validity Checks

STRATEGIES FOR IMPROVING MA SORB CLASSIFICATION Examples from Two States New Jersey Oregon

New Jersey New Jersey: State Generated Actuarial

RRASItems Scoring: Highest possible total score = 111 Low Range: 0 – 36 Moderate Range: 37 – 73 High Range:

Focuses on the current empirical literature to generate items and a scale. Each item is quantified and anchored cutoffs are provided. Method of combining items to generate a score is specified. Levels are tied to specific scores. New Jersey: State Generated Actuarial Advantages

Reliability is an iterative process that takes time to develop. Baserates of scores not initially available. No follow-up data are available. No reoffense probabilities available until prospective study completed. New Jersey: State Generated Actuarial Disadvantages

48 Re-offense Rates by State Risk Levels MN & NJ: 3 Level SystemFL & SC: Offender / Predator (  2 (1) = 3.37, p =.066) (AUCs = , ns) (Zgoba et al., 2014)

STRATEGIES FOR IMPROVING MA SORB CLASSIFICATION Examples from Two States New Jersey Oregon

Oregon: Standard Actuarial

The Static-99R is the chosen risk assessment scale for Oregon, with the following level cutoffs recommended: Level I: Score -3 to 3 (Low) Level II: Score 4 to 5 (Moderate) Level III: Score of 6+ Override and downward departure factors are taken into consideration: Aggravating factors that result in override to a higher level: 1.Deviant Sexual Preference (by STABLE-2007 definition); 2.Emotional Identification with Children (STABLE-2007 definition); 3.High level of psychopathic traits as identified by validated assessment 4.Individual articulates to officials/treatment professional an unwillingness to control future sexually assaultive behaviors and/or plans to reoffend violently or sexually. Mitigating factors that result in downward departure to lower level: 1.Debilitating illness and/or permanent incapacitation years clean record within the community Assessments for aggravating and mitigating factors must be completed by a trained professional.

53 Static 99R Items

Focuses on the current empirical literature to generate items and a scale. Each item is quantified and anchored cutoffs are provided. Method of combining items to generate a score are specified. Levels are tied to specific scores. Oregon: Standard Actuarial Advantages

Extensive follow-up data have been already been gathered. There are existing estimates of the probabilities of recidivism for score levels. Oregon: Standard Actuarial Advantages

Actuarial not made specifically for the local state environment. Tied to standardized instrument that you are less likely to assess for continuous improvement. Disadvantages Oregon: Standard Actuarial

APPLYING THE TWO STRATEGIES TO THE MA SORB CRITERIA

General Issues Creation of separate adult and juvenile actuarials; Creation of separate male and female actuarials; Dealing with the issues of Mental Illness and Intellectual Disabilities. Improving the Current MA SORB Criteria

Strategy 1: NJ Solution Fix the Current MA SORB Criteria for Adult Males Divide instrument into static and dynamic item subsets; Use recent meta-analytic literature to purge items that are not likely to have predictive validity;

Examples of Poor Predictors Released from civil commitment vs. not committed (Knight & Thornton, 2007) Maximum term of incarceration; Current home situation (?vague and unspecified?); Physical condition; Documentation from a licensed mental health professional specifically indicating that offender poses no risk to reoffend;

Recent Threats; Supplemental material; Victim impact statement. Examples of Poor Predictors

Strategy 1: NJ Solution Fix the Current MA SORB Criteria for Adult Males Divide instrument into static and dynamic item subsets; Use recent meta-analytic literature to purge items that are not likely to have predictive validity; Transform remaining items into a quantifiable format with clear cutoffs; Do a small study on a subset of offenders to establish reliability. ? Add items to capture predictive domains not adequately sampled?

Strategy 1: NJ Solution Fix the Current MA SORB Criteria for Adult Males Adjust items with the reliability data; Do a preliminary check on the predictive validity of revised items using existing data bases; Revise items as a function of predictive study and establish preliminary leveling cutoffs; Use the revised instrument, requiring item and total scores of raters for future validation studies.

Strategy 1: NJ Solution Fix the Current MA SORB Criteria for Adult Males Follow all offenders and prospectively assess the instrument’s predictive validity of recidivism; Continually adjust instrument to improve predictive accuracy.

Strategy 2: OR Solution Strategy 2: OR Solution Use the Static99R to determine leveling; Any “aggravating” or “mitigating” criteria should be operationally defined (e.g., STABLE 2007; PCL:R), and its adjustment contribution should be quantitatively specified. SORB has been doing Static99Rs for a while, so use the ones that they have done. Have a team of trained graduate student raters (cheap and accurate) do Static99Rs on remaining offenders.

ESTIMATING LEVEL 3 FREQUENCY

MTC Committed

MTC Not Committed

STATIC-99R Scores (n = 1312) % Zgoba et al., 2014

MA % RSO Level 3 (2010) As cited in Harris, Levenson, & Ackerman, 2012

Moving forward use existing dynamic instruments to create profiles for treatment and management of offenders and for future adjustments. Strategy 2: OR Solution