Business Intelligence and Decision Modeling

Slides:



Advertisements
Similar presentations
Individual level synthetic regression models Session 4.
Advertisements

Statistical Analysis SC504/HS927 Spring Term 2008
Presentation title Date Responder endpoint and continuous endpoint, logistic regression or ANOVA? DSBS 24 OCT 2013 Søren Andersen.
Regression analysis Linear regression Logistic regression.
Logistic Regression.
Week 3. Logistic Regression Overview and applications Additional issues Select Inputs Optimize complexity Transforming Inputs.
1 9. Logistic Regression ECON 251 Research Methods.
Chapter 8 Logistic Regression 1. Introduction Logistic regression extends the ideas of linear regression to the situation where the dependent variable,
SLIDE 1IS 240 – Spring 2010 Logistic Regression The logistic function: The logistic function is useful because it can take as an input any.
Multinomial Logistic Regression
Introduction to Logistic Regression. Simple linear regression Table 1 Age and systolic blood pressure (SBP) among 33 adult women.
EPI 809/Spring Multiple Logistic Regression.
Nemours Biomedical Research Statistics April 23, 2009 Tim Bunnell, Ph.D. & Jobayer Hossain, Ph.D. Nemours Bioinformatics Core Facility.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
1 G Lect 11M Binary outcomes in psychology Can Binary Outcomes Be Studied Using OLS Multiple Regression? Transforming the binary outcome Logistic.
Logistic regression for binary response variables.
A Case Study on Traffic Violations in the City of Colombo Udara Perera Sandun Silva Oshada Senaweera Yogeswaran Akhilan Amani Subawickrama.
MODELS OF QUALITATIVE CHOICE by Bambang Juanda.  Models in which the dependent variable involves two ore more qualitative choices.  Valuable for the.
Occurrence and timing of events depend on Exposure to the risk of an event exposure Risk depends on exposure.
1 G Lect 11W Logistic Regression Review Maximum Likelihood Estimates Probit Regression and Example Model Fit G Multiple Regression Week 11.
Logistic Regression for binary outcomes
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Correlation and Regression SCATTER DIAGRAM The simplest method to assess relationship between two quantitative variables is to draw a scatter diagram.
April 6 Logistic Regression –Estimating probability based on logistic model –Testing differences among multiple groups –Assumptions for model.
Business Intelligence and Decision Modeling Week 11 Predictive Modeling (2) Logistic Regression.
7.4 – Sampling Distribution Statistic: a numerical descriptive measure of a sample Parameter: a numerical descriptive measure of a population.
Limited Dependent Variables Ciaran S. Phibbs May 30, 2012.
Logistic Regression Database Marketing Instructor: N. Kumar.
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 7 Logistic Regression I.
When and why to use Logistic Regression?  The response variable has to be binary or ordinal.  Predictors can be continuous, discrete, or combinations.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Lecture Slide #1 Logistic Regression Analysis Estimation and Interpretation Hypothesis Tests Interpretation Reversing Logits: Probabilities –Averages.
Week 5: Logistic regression analysis Overview Questions from last week What is logistic regression analysis? The mathematical model Interpreting the β.
Assessing Binary Outcomes: Logistic Regression Peter T. Donnan Professor of Epidemiology and Biostatistics Statistics for Health Research.
MBP1010 – Lecture 8: March 1, Odds Ratio/Relative Risk Logistic Regression Survival Analysis Reading: papers on OR and survival analysis (Resources)
1 Chapter 2: Logistic Regression and Correspondence Analysis 2.1 Fitting Ordinal Logistic Regression Models 2.2 Fitting Nominal Logistic Regression Models.
Limited Dependent Variables Ciaran S. Phibbs. Limited Dependent Variables 0-1, small number of options, small counts, etc. 0-1, small number of options,
Warsaw Summer School 2015, OSU Study Abroad Program Advanced Topics: Interaction Logistic Regression.
Multiple Regression  Similar to simple regression, but with more than one independent variable R 2 has same interpretation R 2 has same interpretation.
Logistic Regression. Linear regression – numerical response Logistic regression – binary categorical response eg. has the disease, or unaffected by the.
CHAPTER 10: Logistic Regression. Binary classification Two classes Y = {0,1} Goal is to learn how to correctly classify the input into one of these two.
Heart Disease Example Male residents age Two models examined A) independence 1)logit(╥) = α B) linear logit 1)logit(╥) = α + βx¡
LOGISTIC REGRESSION Binary dependent variable (pass-fail) Odds ratio: p/(1-p) eg. 1/9 means 1 time in 10 pass, 9 times fail Log-odds ratio: y = ln[p/(1-p)]
Logistic regression. Recall the simple linear regression model: y =  0 +  1 x +  where we are trying to predict a continuous dependent variable y from.
Logistic regression (when you have a binary response variable)
Linear Models Binary Logistic Regression. One-Way IVR 2 Hawaiian Bats Examine data.frame on HO Section 1.1 Questions –Is subspecies related to canine.
Binary logistic regression. Characteristic Regression model for target categorized variable explanatory variables – continuous and categorical Estimate.
1 Introduction to Modeling Beyond the Basics (Chapter 7)
Logistic Regression Hal Whitehead BIOL4062/5062.
Logistic Regression and Odds Ratios Psych DeShon.
Statistics and probability Dr. Khaled Ismael Almghari Phone No:
Multiple Regression Analysis Bernhard Kittel Center for Social Science Methodology University of Oldenburg.
Non-Linear Dependent Variables Ciaran S. Phibbs November 17, 2010.
Odds, Logs, and Logits Psych 818 DeShon. Dichotomous Data Assume you have dichotomous outcome data coded as 0 or 1 (true/false, right/wrong) Assume you.
1 BUSI 6220 By Dr. Nick Evangelopoulos, © 2012 Brief overview of Linear Regression Models (Pre-MBA level)
Chapter 13 Logistic regression.
Predicting Brain Size Samantha Stanley Jasmine Dumas Christopher Lee.
Logistic Regression When and why do we use logistic regression?
A priori violations In the following cases, your data violates the normality and homoskedasticity assumption on a priori grounds: (1) count data  Poisson.
Advanced Quantitative Techniques
Logistic Regression.
CHAPTER 10: Logistic Regression
LOGISTIC REGRESSION 1.
Logistic regression One of the most common types of modeling in the biomedical literature Especially case-control studies Used when the outcome is binary.
The Variate COM 631/731.
Introduction to logistic regression a.k.a. Varbrul
females males Analyses with discrete variables
Introduction to Logistic Regression
MPHIL AdvancedEconometrics
Cases. Simple Regression Linear Multiple Regression.
Presentation transcript:

Business Intelligence and Decision Modeling Week 11 Predictive Modeling (2) Logistic Regression

Regressions Simple Regression Multiple Regression Logistic Regression Y = a + bX + e Multiple Regression Y = a + b1X1 + b2X2 + … + bnXn + e Logistic Regression p = 1 / 1+e-z Z = a + b1X1 + b2X2 + …+ bnXn + e

Logistic Regression and Function p = 1 / 1+e-z Where Logit or z = b0+b1x1+b2x2…+bpxp P=1 P=0 Z -6 -4 -2 0 2 4 6

Outline Logistic Regression Purpose Odds and Logit Interpretation

Logistic Regression Models Logistic regression (binary target) Understand risk factors Assumptions same as linear regression Forecasting Split Samples

Logistic Regression: Odds and Probabilities Dichotomised Response (0/1) Response Probability p Non Response Probability (1-p) Odds = p /(1-p) P = odds/(1+odds)

Probabilities and Odss Probability .10 .20 .30 .40 .50 .60 .70 .80 .90 Odds .11 .25 .43 .67 1.00 1.50 2.33 4.00 9.00 Odds = (p / 1-p) P = odds/(1+odds)

Logistic Regression: Logit Logit Calculation Logit  ln(odds) or ln(p/1-p) Inverse Process e(Logit)  odds or p/1-p If P = odds/(1+odds) Then p = e(Logit)/1+e(Logit) Or p = 1 / 1+e-Logit

Logistic Regression and Function p = 1 / 1+e-z Where Logit or z = b0+b1x1+b2x2…+bpxp P=1 P=0 Z -6 -4 -2 0 2 4 6

Example (1) Thus Z=-10.83 + (.28 x age) +(2.30 x gender) Where gender=0 Male et gender =1 Female If Male 40 years old Z = -10.83+(.28 x 40)+(2.30 x 0) Z = .37 Logit e.37 = 1.448 Odds Thus p = 1 / 1+e-z  p = 1 / 1+e-.37 = .59 or P = odds/(1+odds)  p = 1.448/(1+1.448) = .59

Example (2) Z=-10.83 + .28 x age +2.30 x gender If Female 40 years old where gender=0 Male And gender =1 Female If Female 40 years old Z = -10.83+(.28 x 40)+(2.30 x 1) Z = 2.67 e2.67 = 14.44 Odds Logit p = 1 / 1+e-z  p = 1 / 1+e-2.67 = .94 or P = odds/(1+odds)  p = 14.44/(1+14.44) = .93

To Summarize Simple Regression Multiple Regression Logistic Regression Y = a + bX + e Multiple Regression Y = a + b1X1 + b2X2 + … + bnXn + e Logistic Regression p = 1 / 1+e-z Z = a + b1X1 + b2X2 + …+ bnXn + e