Bayesian Linear Regression

Slides:



Advertisements
Similar presentations
Generalised linear mixed models in WinBUGS
Advertisements

Brief introduction on Logistic Regression
Research Support Center Chongming Yang
Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
Logistic Regression Psy 524 Ainsworth.
Chapter 6 Sampling and Sampling Distributions
Copyright © 2010 Pearson Education, Inc. Slide
Inference for Regression
OUTLIER, HETEROSKEDASTICITY,AND NORMALITY
Confidence intervals. Population mean Assumption: sample from normal distribution.
Chapter 7 Sampling and Sampling Distributions
Econ 140 Lecture 191 Heteroskedasticity Lecture 19.
Inference about a Mean Part II
Part III: Inference Topic 6 Sampling and Sampling Distributions
AP Statistics Section 10.2 A CI for Population Mean When is Unknown.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Discriminant Analysis Testing latent variables as predictors of groups.
Basic Analysis of Variance and the General Linear Model Psy 420 Andrew Ainsworth.
1 Overview of Major Statistical Tools UAPP 702 Research Methods for Urban & Public Policy Based on notes by Steven W. Peuquet, Ph.D.
Leedy and Ormrod Ch. 11 Gray Ch. 14
Chapter 12: Analysis of Variance
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Linear Regression Inference
Overview of Major Statistical Tools UAPP 702 Research Methods for Urban & Public Policy Based on notes by Steven W. Peuquet, Ph.D. 1.
CORRELATION & REGRESSION
Inferences for Regression
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Robust Estimators.
Psychology 202a Advanced Psychological Statistics November 12, 2015.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Regression Analysis Intro to OLS Linear Regression.
 Seeks to determine group membership from predictor variables ◦ Given group membership, how many people can we correctly classify?
Chapter 6 Sampling and Sampling Distributions
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
Hierarchical models. Hierarchical with respect to Response being modeled – Outliers – Zeros Parameters in the model – Trends (Us) – Interactions (Bs)
Lecturer: Ing. Martina Hanová, PhD..  How do we evaluate a model?  How do we know if the model we are using is good?  assumptions relate to the (population)
Chapter 12: Correlation and Linear Regression 1.
Applied Regression Analysis BUSI 6220
Linear Regression with One Regression
Statistical Data Analysis - Lecture /04/03
Other confidence intervals
Regression Analysis: Statistical Inference
INF397C Introduction to Research in Information Studies Spring, Day 12
3. The X and Y samples are independent of one another.
Analysis of Covariance (ANCOVA)
Sample Mean Distributions
Chapter 11 Simple Regression
Multiple Regression Analysis
Regression.
Chapter 7: Sampling Distributions
Slides by JOHN LOUCKS St. Edward’s University.
Simple Linear Regression - Introduction
CHAPTER 29: Multiple Regression*
Comparing Several Means: ANOVA
PSY 626: Bayesian Statistics for Psychological Science
More about Normal Distributions
Statistical Assumptions for SLR
Warmup To check the accuracy of a scale, a weight is weighed repeatedly. The scale readings are normally distributed with a standard deviation of
Simple Linear Regression
Quadrat sampling Quadrat shape Quadrat size Lab Regression and ANCOVA
Chapter 7: The Normality Assumption and Inference with OLS
Models, parameters and GLMs
Bayesian Data Analysis in R
Models, parameters and GLMs
Pearson Correlation and R2
Interval Estimation of mean response
F test for Lack of Fit The lack of fit test..
Presentation transcript:

Bayesian Linear Regression

Introducing GLM Important Variable Descriptors: Independent vs Dependent Cardinal vs Ordinal vs Categorical GLM Linear Regression and ANOVA are closely related. In fact, they are both special cases of a more general family of models: General Linear Model (GLM) ANOVA Linear Regression Categorical Predictor Variable(s) Continuous Predictor Variable(s)

OLS Linear Regression Worked Example

Against Causality If a dependent variable depends on an independent variable, that still doesn’t conclusively demonstrate causality nor temporal priority. Height can depend on weight, but doesn’t “happen first” Similarly, even if a linear regression model provides a lot of predictive power, Y might still be causally disconnected from X.

Constructing the Model Once β0 and β1 are known, we construct distributions around the central tendency. μi = β0 + β1*xi The normal distribution varies as follows: yi ~ N(μi, σ)

Constructing the Model OLS assumes homogeneity of variance, or absence of heteroskedasticity. Heteroskedastic Homoskedastic Evenly Distributed Homoskedastic Bimodally Distributed

Using Student Distribution The Student-t Distribution generalizes the Normal Distribution y ~ T(μ, σ, v) As v (degrees of freedom) approaches infinity, it approximates the Normal T(μ, σ, v = ∞) = Ɲ(μ, σ) We will here use the T distribution instead of the Normal Distribution to more easily accommodate outliers (heavier tails)

Bayesian Model The core of OLS Linear Regression is setting μi = β0 + β1*xi With this hierarchical model, we must shmear probability mass across the space of four parameters: β0, β1, σ, v To discover how these parameters update on the data, we use MCMC to approximate Bayesian inference:

Bayesian Model To import our model into JAGs, we simply transcribe our graphic into code: The “z” prefix denotes standardized data, explained later

Data Normalization In frequentist OLS, it is considered best practice to mean-center your IVs: X’ = X - μx For Bayesian Inference, we will go a bit further & standardize: X’ = (X - μx) / σx What’s the point? Mean Centering: Decorrelates parameters Normalization: priors less sensitive to scale of the data

Where To Normalize? Recall division of labor between R & JAGS. Graphics MCMC Diagnostics Normalization in R is trivial enough. However, JAGS also provides a data manipulation interface. Benefit of normalizing in JAGS: Diagnostics are easier to consume Example

Bayesian Linear Regression JAGS Implementation