Copula Regression By Rahul A. Parsa Drake University &

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Pattern Recognition and Machine Learning
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Capacity of MIMO Channels: Asymptotic Evaluation Under Correlated Fading Presented by: Zhou Yuan University of Houston 10/22/2009.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
The Simple Linear Regression Model: Specification and Estimation
Factor Analysis Purpose of Factor Analysis
GRA 6020 Multivariate Statistics The regression model OLS Regression Ulf H. Olsson Professor of Statistics.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Statistics 350 Lecture 14. Today Last Day: Matrix results and Chapter 5 Today: More matrix results and Chapter 5 Please read Chapter 5.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
Presenting: Assaf Tzabari
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Mixed models Various types of models and their relation
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Linear and generalised linear models
Violations of Assumptions In Least Squares Regression.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Maximum likelihood (ML)
Lecture II-2: Probability Review
Modern Navigation Thomas Herring
Review of Lecture Two Linear Regression Normal Equation
Moment Generating Functions 1/33. Contents Review of Continuous Distribution Functions 2/33.
Moment Generating Functions
Traffic Modeling.
Chapter 14 Monte Carlo Simulation Introduction Find several parameters Parameter follow the specific probability distribution Generate parameter.
Elements of Financial Risk Management Second Edition © 2012 by Peter Christoffersen 1 Distributions and Copulas for Integrated Risk Management Elements.
MULTIPLE TRIANGLE MODELLING ( or MPTF ) APPLICATIONS MULTIPLE LINES OF BUSINESS- DIVERSIFICATION? MULTIPLE SEGMENTS –MEDICAL VERSUS INDEMNITY –SAME LINE,
1 Statistical Distribution Fitting Dr. Jason Merrick.
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
Danila Filipponi Simonetta Cozzi ISTAT, Italy Outlier Identification Procedures for Contingency Tables in Longitudinal Data Roma,8-11 July 2008.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
1 GLM I: Introduction to Generalized Linear Models By Curtis Gary Dean Distinguished Professor of Actuarial Science Ball State University By Curtis Gary.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Generalized Linear Models (GLMs) and Their Applications.
Lecture 2: Statistical learning primer for biologists
Estimation of covariance matrix under informative sampling Julia Aru University of Tartu and Statistics Estonia Tartu, June 25-29, 2007.
Topic 5: Continuous Random Variables and Probability Distributions CEE 11 Spring 2002 Dr. Amelia Regan These notes draw liberally from the class text,
Chapter 20 Statistical Considerations Lecture Slides The McGraw-Hill Companies © 2012.
Review of statistical modeling and probability theory Alan Moses ML4bio.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
MathematicalMarketing Slide 5.1 OLS Chapter 5: Ordinary Least Square Regression We will be discussing  The Linear Regression Model  Estimation of the.
Computacion Inteligente Least-Square Methods for System Identification.
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Usman Roshan CS 675 Machine Learning
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
The regression model in matrix form
EC 331 The Theory of and applications of Maximum Likelihood Method
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
OVERVIEW OF LINEAR MODELS
OVERVIEW OF LINEAR MODELS
Chapter 14 Monte Carlo Simulation
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Chapter 3 : Random Variables
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
Topic 11: Matrix Approach to Linear Regression
Probabilistic Surrogate Models
Presentation transcript:

Copula Regression By Rahul A. Parsa Drake University & Stuart A. Klugman Society of Actuaries

Outline of Talk OLS Regression Generalized Linear Models (GLM) Copula Regression Continuous case Discrete Case Examples

Notation Notation: Y – Dependent Variable Assumption Y is related to X’s in some functional form

OLS Regression Y is linearly related to X’s OLS Model

OLS Regression Estimated Model

OLS Multivariate Normal Distribution Assume Jointly follow a multivariate normal distribution Then the conditional distribution of Y | X follows normal distribution with mean and variance given by

OLS & MVN Y-hat = Estimated Conditional mean It is the MLE Estimated Conditional Variance is the error variance OLS and MLE result in same values Closed form solution exists

GLM Y belongs to an exponential family of distributions g is called the link function x's are not random Y|x belongs to the exponential family Conditional variance is no longer constant Parameters are estimated by MLE using numerical methods

GLM Generalization of GLM: Y can be any distribution (See Loss Models) Computing predicted values is difficult No convenient expression conditional variance

Copula Regression Y can have any distribution Each Xi can have any distribution The joint distribution is described by a Copula Estimate Y by E(Y|X=x) – conditional mean

Copula Ideal Copulas will have the following properties: ease of simulation closed form for conditional density different degrees of association available for different pairs of variables.  Good Candidates are: Gaussian or MVN Copula t-Copula

MVN Copula CDF for MVN is Copula is Where G is the multivariate normal cdf with zero mean, unit variance, and correlation matrix R. Density of MVN Copula is Where v is a vector with ith element

Conditional Distribution in MVN Copula The conditional distribution of xn given x1 ….xn-1 is Where

Copula Regression Continuous Case Parameters are estimated by MLE. If are continuous variables, then we use previous equation to find the conditional mean. one-dimensional numerical integration is needed to compute the mean.

Copula Regression Discrete Case When one of the covariates is discrete Problem: determining discrete probabilities from the Gaussian copula requires computing many multivariate normal distribution function values and thus computing the likelihood function is difficult Solution: Replace discrete distribution by a continuous distribution using a uniform kernel.

Copula Regression – Standard Errors How to compute standard errors of the estimates? As n -> ∞, MLE , converges to a normal distribution with mean q and variance I(q)-1, where I(q) – Information Matrix.

How to compute Standard Errors Loss Models: “To obtain information matrix, it is necessary to take both derivatives and expected values, which is not always easy. A way to avoid this problem is to simply not take the expected value.” It is called “Observed Information.”

Examples All examples have three variables R Matrix : Error measured by Also compared to OLS 1 0.7

Example 1 Dependent – X3 - Gamma Though X2 is simulated from Pareto, parameter estimates do not converge, gamma model fit Error: Variables X1-Pareto X2-Pareto X3-Gamma Parameters 3, 100 4, 300 MLE 3.44, 161.11 1.04, 112.003 3.77, 85.93 Copula 59000.5 OLS 637172.8

Ex 1 - Standard Errors Diagonal terms are standard deviations and off-diagonal terms are correlations

Example 1 - Cont Maximum likelihood Estimate of Correlation Matrix 1 0.711 0.699 0.713 R-hat =

Example 2 Dependent – X3 - Gamma X1 & X2 estimated Empirically Error: Variables X1-Pareto X2-Pareto X3-Gamma Parameters 3, 100 4, 300 MLE F(x) = x/n – 1/2n f(x) = 1/n 4.03, 81.04 Copula 595,947.5 OLS 637,172.8 GLM 814,264.754

Example 3 Dependent – X3 - Gamma Pareto for X2 estimated by Exponential Error: Variables X1-Poisson X2-Pareto X3-Gamma Parameters 5 4, 300 3, 100 MLE 5.65 119.39 3.67, 88.98 Copula 574,968 OLS 582,459.5

Example 4 Dependent – X3 - Gamma X1 & X2 estimated Empirically C = # of obs ≤ x and a = (# of obs = x) Error: Variables X1-Poisson X2-Pareto X3-Gamma Parameters 5 4, 300 3, 100 MLE F(x) = c/n + a/2n f(x) = a/n F(x) = x/n – 1/2n f(x) = 1/n 3.96, 82.48 Copula OLS GLM 559,888.8 582,459.5 652,708.98

Example 5 Dependent – X1 - Poisson X2, estimated by Exponential Error: Variables X1-Poisson X2-Pareto X3-Gamma Parameters 5 4, 300 3, 100 MLE 5.65 119.39 3.66, 88.98 Copula 108.97 OLS 114.66

Example 6 Dependent – X1 - Poisson X2 & X3 estimated by Empirically Error: Variables X1-Poisson X2-Pareto X3-Gamma Parameters 5 4, 300 3, 100 MLE 5.67 F(x) = x/n – 1/2n f(x) = 1/n Copula 110.04 OLS 114.66