Presentation is loading. Please wait.

Presentation is loading. Please wait.

Creating Empirical Models Constructing a Simple Correlation and Regression-based Forecast Model Christopher Oludhe, Department of Meteorology, University.

Similar presentations


Presentation on theme: "Creating Empirical Models Constructing a Simple Correlation and Regression-based Forecast Model Christopher Oludhe, Department of Meteorology, University."— Presentation transcript:

1 Creating Empirical Models Constructing a Simple Correlation and Regression-based Forecast Model Christopher Oludhe, Department of Meteorology, University of Nairobi Clips training workshop for Eastern and Southern Africa, DMCN. 30 th July 2002

2 Simple Linear Correlation Analysis Many problems in seasonal climate prediction start by trying to establish some relationship (linear) between two sets of variables. An example would be to try and see whether the (SST) over any of the global oceans (variable one) is related to rainfall (variable two) at a certain given location of the globe.

3 Simple Linear Correlation Cont.. The Knowledge of such a relationship would be useful in that the expected rainfall of the given location can be predicted if the SSTs of the global Oceans are known in advance. The strength of the two relationship can be determined by computing the Pearson’s coefficient of correlation, r.

4 Simple Linear Correlation Cont.. The statistical significance of the computed correlation coefficient r may be tested by using the t- statistic given by: Accept or reject your null hypothesis ( r = 0) depending on the results of the comparison between computed and tabulated t.

5 Simple Linear Regression Equation A simple linear regression equation gives the functional relationship between two variables such as: Y =  +  x where x is the independent variable (predictor) and Y the dependent variable (response or predictand). The estimation of the regression constants,  (Y- intercept) and  (slope of the line), are possible through the method of least-squares.

6 Regression Cont.. The solutions for the regression constants are given by the relations: and

7 Plotting of paired data Given a set of paired standardised rainfall (Y ) and SST (X ) data below i x i y i i x i y i 1-1.3719-0.217911 0.0670 0.4004 2-0.4962 0.913312 0.0131-1.0518 3 0.9218-0.475313 1.7473-1.2373 4-1.8650 2.025414 0.3468-1.0691 5-1.0855 0.418615-0.2330-0.4616 6-1.8062 2.125516 0.0106-0.6662 7 0.0720 0.357717 1.0150-0.9299 8 0.5173-0.010618 1.5223 1.5044 9 0.2192-0.823619-0.7204 0.2140 10 0.7199-0.549820 0.4061-0.4662

8 Scatter Plot and Line of Best Fit

9 Linear Regression Fit It can be seen that the relationship is linear but negatively, i.e. when the SST index increases (decreases), the rainfall index decreases (increases), i.e, positive SST index are associated with a negative rainfall index, or drier than average conditions. Using this type of relationship, it is possible to make a qualitative statement regarding the expected rainfall for a coming season if knowledge of the seasonal lag SST index can obtained just before the beginning of the season to be forecasted.

10 Goodness of fit measure The goodness of “fit” of a regression model can be determined by examining the mean-squared error (MSE) in the ANOVA table output. This measure indicates the variability of the observed values around the forecast regression line. A perfect linear relationship between the predictor and predictand gives an MSE of zero, while poor fits results in large values of MSE. Another measure of the fit of a regression is the coefficient of determination (R 2 ) which is, the squared value of the Pearson correlation coefficient between predictor and predictand.

11 Measure Cont.. Qualitatively, R 2 can be interpreted as the proportion of the variance of the predictand that is described or accounted for by the regression. For a perfect regression, the R 2 = 1, while for R 2 close to 0 indicates that very little of the variance is being explained by the regression line. In majority of applications, however, the response of a predictand can be predicted more adequately by a collection of many variables and not just on the basis of a single independent input variable.

12 Multiple Linear Regression In a multiple linear regression model, a single predictant, Y, (e.g. SOND rainfall) has more than one predictor variable, i.e, it can be influenced by ENSO, QBO, SSTs over the Indian Ocean AND/OR the Atlantic Ocean, etc. For K predictors: Y = β 0 + β 1 x 1 + β 2 x 2 + … + β k x k The procedure for estimating the regression coefficients is the same as those for simple linear regression models.

13 Stepwise Regression analysis Forward Selection: In this procedure, only the best potential predictors that improves the model the most, are examined individually and added into the model equation, starting with the one that explains the highest variance, etc. Backward Elimination: The regression model starts with all potential predictors and at each step of model construction, the least important predictor is removed until only the best predictors remain. A stopping criteria should be selected in both cases.

14 Cross-Validation Year1Year 2Year 3Year 4Year 5Year 6 Model 1 omitted Model 2 omitted Model 3 omitted Model 4 omitted Model 5 omitted Model 6 omitted Model 7

15 Forecast Skill Estimation (Contingency Table) OAOA ONON OBOB FAFA RST FNFN UVW FBFB XYZ O: Observed F: Forecast A: Above-normal N: Near-normal B: Below-normal

16 Accuracy Measures of Multicategory Forecasts (1) Hit Score (HS): Number of times a correct category is forecast HS = R+V+Z

17 Accuracy Measures of Multicategory Forecasts (2) False Alarm Ratio (FAR): The fraction of forecast events that failed to materialize Best FAR=0; worst FAR=1 For Above-Normal=(S+T)/(R+S+T) For Near-Normal=(U+W)/(U+V+W) For Below-Normal=(X+Y)/(X+Y+Z)

18 Accuracy Measures of Multicategory Forecasts (3) Bias: Comparison of the average forecast with the average observation Bias > 1 : overforecasting Bias < 1 : underforecasting For Above-Normal=(R+S+T)/(R+U+X) For Near-Normal=(U+V+W)/(S+V+Y) For Below-Normal=(X+Y+Z)/(T+W+Z)

19 Example of Model Testing (1) Regression lines for some of the 20 cross-validation models: Mod. 1: Y 1 = 0.0579-0.6434x (years 2 to 20) Mod. 2: Y 2 =-0.0339-0.5421x (year 1 and 3 to 20) : Mod. 18: Y 18 =-0.1423-0.7878x (years 1 to 17 and 19 and 20) : Mod. 20: Y 20 = 0.0127-0.5543x (years 1 to 19)

20 Example of Model Testing (2) Linear Fits of the 20 cross-validation models. The red line is the fit of Model 18 which excludes the outlier at about (1.5,1.5)

21 Example of Model Testing (3) Cross-val. forecasts (dashed) and observed (solid) using data from 18 seasons. Horizontal lines on either side of zero line: upper and lower limits of Near-Normal

22 Example of Model Testing (4) OAOA ONON OBOB FAFA 410 FNFN 224 FBFB 032 HS=4+2+2 BIAS A =(4+1+0)/(4+2+0); BIAS N =(2+2+4)/(1+2+3); BIAS B =(0+3+2)/(0+4+2) FAR A =(1+0)/(4+1+0); FAR N =(2+4)/(2+2+4); FAR B =(0+3)/(0+3+2)


Download ppt "Creating Empirical Models Constructing a Simple Correlation and Regression-based Forecast Model Christopher Oludhe, Department of Meteorology, University."

Similar presentations


Ads by Google