Using evidence-based decision trees instead of formulas to identify at-risk readers Sharon Koon Yaacov Petscher Barbara R. Foorman Florida Center For Reading.

Slides:



Advertisements
Similar presentations
Continued Psy 524 Ainsworth
Advertisements

Random Forest Predrag Radenković 3237/10
Brief introduction on Logistic Regression
Logistic Regression Psy 524 Ainsworth.
Chapter 10 Decision Making © 2013 by Nelson Education.
Week 3. Logistic Regression Overview and applications Additional issues Select Inputs Optimize complexity Transforming Inputs.
Chapter 8 – Logistic Regression
Departments of Medicine and Biostatistics
Chapter 7 – Classification and Regression Trees
Chapter 7 – Classification and Regression Trees
Regression With Categorical Variables. Overview Regression with Categorical Predictors Logistic Regression.
Practical Meta-Analysis -- D. B. Wilson
Multiple Regression – Basic Relationships
Chapter 7 Correlational Research Gay, Mills, and Airasian
Multinomial Logistic Regression Basic Relationships
Chapter 14 Inferential Data Analysis
Ensemble Learning (2), Tree and Forest
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Split Sample Validation General criteria for split sample validation Sample problems.
SW388R7 Data Analysis & Computers II Slide 1 Logistic Regression – Hierarchical Entry of Variables Sample Problem Steps in Solving Problems.
Wednesday PM  Presentation of AM results  Multiple linear regression Simultaneous Simultaneous Stepwise Stepwise Hierarchical Hierarchical  Logistic.
ICSD District RtI Committee Agenda 3/13/12 3:45- Review of Our Norms and today’s agenda 4:00- Defining RtI and screening tool criteria 4:30- Begin review.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 12 Describing Data.
Response to Intervention (RtI) practices in educational settings aim to identify students who are at risk for academic struggles. Typically, RtI tries.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Hierarchical Binary Logistic Regression
Biostatistics Case Studies 2005 Peter D. Christenson Biostatistician Session 5: Classification Trees: An Alternative to Logistic.
Statistical Fundamentals: Using Microsoft Excel for Univariate and Bivariate Analysis Alfred P. Rovai Hypothesis Testing PowerPoint Prepared by Alfred.
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
Multinomial Logistic Regression Basic Relationships
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Chapter 9 – Classification and Regression Trees
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Slide 1 The SPSS Sample Problem To demonstrate these concepts, we will work the sample problem for logistic regression in SPSS Professional Statistics.
Patterns of Event Causality Suggest More Effective Corrective Actions Abstract: The Occurrence Reporting and Processing System (ORPS) has used a consistent.
SW388R7 Data Analysis & Computers II Slide 1 Logistic Regression – Hierarchical Entry of Variables Sample Problem Steps in Solving Problems Homework Problems.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Introduction to Quantitative Research Analysis and SPSS SW242 – Session 6 Slides.
2006 OSEP Project Directors Meeting 1 Screening and Progress Monitoring for Identification of Reading Disabilities within an RTI Model Screening and Progress.
Chapter 4: Introduction to Predictive Modeling: Regressions
Adjusted from slides attributed to Andrew Ainsworth
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Regression Analysis: Part 2 Inference Dummies / Interactions Multicollinearity / Heteroscedasticity Residual Analysis / Outliers.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Logistic Regression Analysis Gerrit Rooks
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Two-Group Discriminant Function Analysis. Overview You wish to predict group membership. There are only two groups. Your predictor variables are continuous.
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
(Slides not created solely by me – the internet is a wonderful tool) SW388R7 Data Analysis & Compute rs II Slide 1.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Barbara Foorman, Yaacov Petscher, & Chris Schatschneider, Florida Center for Reading Research, Florida State University.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Logistic Regression: Regression with a Binary Dependent Variable.
Lecture Slides Elementary Statistics Twelfth Edition
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
BINARY LOGISTIC REGRESSION
JMP Discovery Summit 2016 Janet Alvarado
Logistic Regression APKC – STATS AFAC (2016).
Regression Analysis Part D Model Building
Diagnosis II Dr. Brent E. Faught, Ph.D. Assistant Professor
Multiple logistic regression
CLAS/RSF Meeting Dr. Barbara Foorman, Mr. Kevin Smith and Ms. Laurie Lee Regional Educational Laboratory (REL) Southeast at Florida State University.
Multiple Regression – Split Sample Validation
Chapter 6 Logistic Regression: Regression with a Binary Dependent Variable Copyright © 2010 Pearson Education, Inc., publishing as Prentice-Hall.
Dr. Phyllis Underwood REL Southeast
Developing Honors College Admissions Rubric to Ensure Student Success
Presentation transcript:

Using evidence-based decision trees instead of formulas to identify at-risk readers Sharon Koon Yaacov Petscher Barbara R. Foorman Florida Center For Reading Research

This information is being provided as part of a presentation administered by the Regional Educational Laboratory Southeast. Information and materials mentioned or shown during this presentation are provided as resources and examples for the viewer's convenience. Their inclusion is not intended as an endorsement by the Regional Educational Laboratory Southeast or its funding source, the Institute of Education Sciences (Contract ED-IES-12-C-0011). In addition, the instructional practices and assessments discussed or shown in these presentations are not intended to mandate, direct, or control a State’s, local educational agency’s, or school’s specific instructional content, academic achievement system and assessments, curriculum, or program of instruction. State and local programs may use any instructional content, achievement system and assessments, curriculum, or program of instruction they wish. 2

REL Southeast

Overview Educators need to understand how students are identified as at risk for reading problems. Early warning systems (EWS) — or diagnostic systems — can be used for this purpose. EWS provide opportunities for interventions to occur that may prevent an anticipated negative outcome (e.g. retention, dropping out). Methods for developing EWS include logistic regression (LR) and classification and regression tree (CART) analysis. While the CART model is an emerging tool in the field of education, limited research exists on its comparability to LR when using multivariate assessments of reading to screen for reading difficulties.

Research question This study used data from a sample of students in grades 1 and 2 in Florida public schools during the 2012/13 academic year to answer the following research question: How do CART models compare with logistic regression models in predicting poor performance on the reading comprehension subtest of the Stanford Achievement Test?

Logistic Regression (LR)

LR (continued) Log-odds for the likelihood of achieving one of the two categories of the selected outcome (e.g., pass/fail a test) can be transformed into probabilities. When there are one or two predictors, contingency tables present a simple way to display the classification rules based on probabilities. Test 1 Test

Sample 2 x 2 classification table Screening Assessment Outcome Assessment FailPass At riskA: True positiveB: False positive Not at riskC: False negativeD: True negative LR (continued) Sensitivity (the proportion of true positives; A/(A+C)]. Specificity (the proportion of true negatives; D/(D+B). Positive predictive power [the proportion of students who are identified as at-risk on the screening assessment and fail the outcome assessment; A/(A+B)] Negative predictive power [the proportion of students identified as not at-risk on the screening assessment and pass the outcome assessment; D/(C+D)]. Measures of classification accuracy

LR (continued) Primary Advantages Widely accepted approach to predicting future performance on similar samples of students. Offers tests of statistical significance of the predictors. Primary Disadvantage Regression equations are often difficult for practitioners to understand and apply.

Classification and Regression Tree (CART) CART uses a set of “if-then” clauses, in the form of a decision tree, to classify individuals, which can easily be applied by practitioners. Example: For two measures which both have score ranges of , it may be observed that students who score less than 244 on Test 1 and less than 350 on test 2 are identified as at-risk.

CART (continued) The CART analysis searches for the optimal split on the predictor variables which best separates the sample into mutually exclusive subsamples called nodes. A variable can appear many times. The nodes form either a rectangular box (terminal node) or a circle (nonterminal node). Circles split again when there is a difference between students on the predictor variables or a stopping rule has not been reached (e.g., minimum split size).

CART (continued) Model specifications: Minimum split—specifies the minimum number of cases that must exist in a node in order for a split to be attempted. Minimum complexity parameter (cp)—specifies the minimum decrease in the overall lack of fit that must result by an additional split. Fit is measured by the model’s relative error, which is equivalent to 1-R 2 (Steinberg, 2013). Loss matrix—is used to weight classification errors differently. To increase the negative predictive power, false negatives would be specified as more costly. The default is equal weight.

CART (continued) CART models are predominantly found in medical applications (i.e., diagnosis and prognosis of illnesses) because of the method’s suitability in generating clinical decision rules. While the CART model is more frequently found in medical literature, it has begun to emerge in educational research.

CART (continued) Primary Advantages Operates on the original test scores’ scale (instead of odds-, log-odds or predicted probabilities). Results can be communicated using decision trees. Not sensitive to the presence of the outliers in the data or collinearity between variables. Models complex interactions among predictors which may be difficult to detect in the regression framework. Primary Disadvantage Complex trees can be difficult to interpret.

Data Large school district in Florida Students were tested on both the Florida Assessment for Instruction in Reading – Florida Standards (FAIR-FS) and Stanford Achievement Test Series, Tenth Edition (SAT-10) in 2012/13 as a part of a linking study conducted under a Reading for Understanding assessment grant. 986 students in grade 1 and 887 students in grade 2 from this archival dataset were included in this analysis.

Measures: Predictors FAIR-FS : – word reading (grades 1 and 2) – word building (grade 1) – spelling (grade 2) – vocabulary pairs (grades 1 and 2) – following directions (grades 1 and 2) Developmental scale scores range from 200 to 800, with a mean of 500 and a standard deviation of 100.

Measures: Outcome SAT-10 – reading comprehension subtest – Percentile scores on the SAT-10 were dichotomized so that scores at or above the 40th percentile were coded as 1 for “not at risk” and scores below the 40th percentile were coded as 0 for “at risk.”

Preliminary steps Deletion of univariate and multivariate outliers Data imputation (less than 10% missing in each grade) – Mean of 20,000 imputed values was used (CART needed one imputed file) Grade file split into calibration (80%) and validation (20%) data sets. – Calibration = “training” data set – Validation = “testing” data set

Grade 1 correlations Grade 1 correlations between measures (n=780) SAT-101 2Following Directions.41 ** 1 3Vocabulary Pairs.52 **.51 ** 1 4Word Reading.64 **.36 **.53 ** 1 5Word Building.53 **.45 **.51 **.61 ** 1 **. Correlation is significant at the 0.01 level (2-tailed).

Grade 2 correlations Grade 2 correlations between measures (n=706) SAT-101 2Following Directions.40 ** 1 3Spelling.60 **.33 ** 1 4Vocabulary Pairs.48 **.40 **.44 ** 1 5Word Reading.62 **.36 **.78 **.46 ** 1 **. Correlation is significant at the 0.01 level (2-tailed).

LR steps Based on the correlations between the individual FAIR-FS tests and the dichotomized SAT-10 variable, the FAIR-FS test scores were entered into the logistic regression ordered by correlational magnitude. FAIR-FS tests which added at least 2% unique variance above the test already in the model, as measured by the Nagelkerke pseudo R-squared, were retained for the final classification model. Analyses conducted using SPSS Statistics 21

CART steps CART models assessed the individual performance of each FAIR-FS test, at every available cut point, in classifying students into risk and no-risk categories. Several specifications were used to limit the number of splits, including: – minimal split size = 3 – complexity parameter (model specific; identified using cross validation statistics; grade 1 =.02; grade 2 =.016). The intention was to build and prune trees based on maximizing the negative predictive power of.85. To accomplish this, revisions to the model in grade 2 included the specification of a loss matrix. Analyses conducted using R rpart package

CART steps (cp values) Grade 1 Grade 2 A recommended minimum standard is the value of the complexity parameter which results in a cross validation relative error less than one standard error above the minimum cross validation relative error (Therneau, Atkinson, & Ripley, 2013).

Logistic Regression Results Logit = *(word reading score) *(word building score) *(vocabulary pairs score) (following directions was dropped from the model, ΔR 2 <.02) Nagelkerke pseudo R 2 =.72 Logits are then transformed to probabilities and then probabilities are used to establish risk classification (e.g., <.5 may be considered at risk). Grade 1 Variable importance: word reading, vocabulary pairs, word building, following directions R 2 =.64 CART Results

Logistic Regression Results Logit = *(word reading score) *(spelling score) *(vocabulary pairs score) *(following directions score) Nagelkerke pseudo R 2 =.70 Logits are then transformed to probabilities and then probabilities are used to establish risk classification (e.g., <.5 may be considered at risk) Grade 2 Variable importance: word reading, spelling, vocabulary pairs, following directions. R 2 =.71 CART Results

Summary by model Validation results (n=206) ModelSensitivitySpecificity Positive Predictive Power Negative Predictive Power Overall Proportion Correct Grade 1 CART Logistic regression Grade 2 CART Logistic regression

Implications The CART results were found to be comparable to those of logistic regression, with the results of both methods yielding negative predictive power greater than the recommended standard of.90. Given the comparability, the CART model may be more appealing in an educational context due to the ease in which the results can be communicated to practitioners.

Limitations The specifications used in this study were designed to meet or exceed negative predictive power of 0.85, while maintaining acceptable levels of sensitivity and specificity. Using a different measure of classification accuracy may affect the results in this report by favoring one method over the other.

Future research The use of CART models in EWS could be studied to determine whether school staff find them easier to use than logistic regression. As noted in the literature review, both CART and LR may be used in a complimentary approach, since each method presents both advantages and disadvantages and provides different tools to the researcher.

Thank you! Website: rel-se.fsu.edu