Download presentation
Presentation is loading. Please wait.
1
6-4 Other Aspects of Regression
6-4.1 Polynomial Models
2
6-4 Other Aspects of Regression
6-4.1 Polynomial Models
3
6-4 Other Aspects of Regression
6-4.1 Polynomial Models Suppose that we wanted to test the contribution of the second-order terms to this model. In other words, what is the value of expanding the model to include the additional terms? ๐= ๐ฝ 0 + ๐ฝ 1 ๐โ ๐ฝ 2 ๐
โ ๐
4
6-4 Other Aspects of Regression
6-4.1 Polynomial Models Full Model: Reduced Model: ๐= ๐ฝ 0 + ๐ฝ 1 ๐โ ๐ฝ 2 ๐
โ ๐ ๐ป 0 : ๐ฝ 12 = ๐ฝ 11 = ๐ฝ 22 = 0 ๐ป 1 :๐ด๐ก ๐๐๐๐ ๐ก ๐๐๐ ๐๐ ๐กโ๐ ๐ฝ โฒ ๐ โ 0 ๐ 0 = (170.73โ11.37) (5โ2) (16โ6) = 46.72 Tabled F = f0.05;3,10 = 2.44 (p-value < )
5
6-4 Other Aspects of Regression
Example 6-9 OPTIONS NOOVP NODATE NONUMBER; DATA ex69; INPUT YIELD TEMP RATIO; TEMPC=TEMP ; RATIOC=RATIO ; TEMRATC=TEMPC*RATIOC; TEMPCSQ=TEMPC**2; RATIOCSQ=RATIOC**2; CARDS; PROC REG DATA=EX69; MODEL YIELD= TEMPC RATIOC TEMRATC TEMPCSQ RATIOCSQ/VIF; TITLE 'QUADRATIC REGRESSION MODEL - FULL MODEL'; MODEL YIELD=TEMPC RATIOC/VIF; TITLE 'LINEAR REGRESSION MODEL - REDUCED MODEL'; RUN; QUIT;
6
6-4 Other Aspects of Regression
7
6-4 Other Aspects of Regression
8
6-4 Other Aspects of Regression
9
6-4 Other Aspects of Regression
10
6-4 Other Aspects of Regression
11
6-4 Other Aspects of Regression
12
6-4 Other Aspects of Regression
Residual Plots (b) The variance of the observations may by increasing with time or with the magnitude of yi or xi. Data transformation on the response y is often used to eliminate this problem ( ๐ฆ , ln ๐ฆ, 1 ๐ฆ ). (c) Plots of residuals against ๐ฆ ๐ and xi also indicate inequality of variance. (d) Indicates model inadequacy; that is, higher-order terms should be added to the model, a transformation on the x-variable or the y-variable (or both) should be considered, or other regressors should be considered (quadratic or exponential model)
13
6-4 Other Aspects of Regression
Example OPTIONS NOOVP NODATE NONUMBER; DATA BIDS; INFILE 'C:\Users\korea\Desktop\Working Folder 2017\imen214-stats\ch06\data\bids.dat'; INPUT PRICE QUANTITY BIDS; LOGPRICE=LOG(PRICE); RECPRICE=1/PRICE; QUANSQ=QUANTITY**2; ODS GRAPHICS ON; proc sgplot; scatter x= quantity y=price; TITLE 'Scatter Plot of PRICE vs. QUANTITY'; PROC REG DATA=BIDS; MODEL PRICE= QUANTITY; TITLE 'LINEAR REGRESSION OF PRICE VS. QUANTITY'; MODEL LOGPRICE= QUANTITY; TITLE 'LINEAR REGRESSION OF LOGPRICE VS. QUANTITY'; MODEL RECPRICE= QUANTITY; TITLE 'LINEAR REGRESSION OF RECPRICE VS. QUANTITY'; MODEL PRICE= QUANTITY QUANSQ; TITLE 'QUADRATIC REGRESSION OF PRICE VS. QUANTITY'; RUN; ods graphics off; QUIT; 153.32 1 4 74.11 7.2 10 29.72 16.7 5 54.67 11.9 68.39 9.3 119.04 3.7 116.14 1.7 6 146.49 0.1 9 81.81 7.8 19.58 18.4 141.08 2.9 101.72 4.7 24.88 17.4 19.43 39.63 11.2 151.13 1.6 7 79.18 7.3 204.94 0.2 81.06 6.8 37.62 11.4 8 17.13 20 3 37.81 13.4 130.72 1.8 2 26.07 18.5 39.59 14.7 66.2 9.1
14
6-4 Other Aspects of Regression
15
6-4 Other Aspects of Regression
16
6-4 Other Aspects of Regression
17
6-4 Other Aspects of Regression
18
6-4 Other Aspects of Regression
19
6-4 Other Aspects of Regression
20
6-4 Other Aspects of Regression
21
6-4 Other Aspects of Regression
22
6-4 Other Aspects of Regression
23
6-4 Other Aspects of Regression
24
6-4 Other Aspects of Regression
25
6-4 Other Aspects of Regression
26
6-4 Other Aspects of Regression
27
6-4 Other Aspects of Regression
28
6-4 Other Aspects of Regression
29
6-4 Other Aspects of Regression
30
6-4 Other Aspects of Regression
6-4.2 Categorical Regressors Many problems may involve qualitative or categorical variables. The usual method for the different levels of a qualitative variable is to use indicator variables. For example, to introduce the effect of two different operators into a regression model, we could define an indicator variable as follows:
31
6-4 Other Aspects of Regression
Example 6-10 Y=gas mileage, x1=engine displacement, x2=horse power x3=0 if automatic transmission 1 if manual transmission ๐= ๐ฝ 0 + ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 + ๐ฝ 3 ๐ฅ 3 +๐ if automatic (x3=0), then ๐= ๐ฝ 0 + ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 +๐ if manual (x3=1), then ๐= ๐ฝ 0 + ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 + ๐ฝ 3 +๐ ๐=( ๐ฝ 0 + ๐ฝ 3 )+ ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 +๐ It is unreasonable because x1, x2 effects to x3 are not involved in the model Interaction model: ๐= ๐ฝ 0 + ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 + ๐ฝ 3 ๐ฅ 3 + ๐ฝ 13 ๐ฅ 1 ๐ฅ 3 + ๐ฝ 23 ๐ฅ 2 ๐ฅ 3 +๐ ๐= ๐ฝ 0 + ๐ฝ 1 ๐ฅ 1 + ๐ฝ 2 ๐ฅ 2 + ๐ฝ 3 ๐ฅ 3 + ๐ฝ 13 ๐ฅ 1 + ๐ฝ 23 ๐ฅ 2 +๐ ๐= ๐ฝ 0 + ๐ฝ ๐ฝ 1 + ๐ฝ 13 ๐ฅ 1 + ๐ฝ 2 + ๐ฝ 23 ๐ฅ 2 +๐
32
6-4 Other Aspects of Regression
Dummy Variables Many times a qualitative variable seems to be needed in a regression model. This can be accomplished by creating dummy variables or indicator variables. If a qualitative variable has ๐ levels you will need ๐โ1 dummy variables. Notice that in ANOVA if a treatment had ๐ levels it had ๐โ1 degrees of freedom. The ith dummy variable is defined as ๐ ๐ = ๐๐ ๐๐ ๐๐กโ ๐๐๐ฃ๐๐ ๐๐ ๐๐ข๐๐๐๐ก๐๐ก๐๐ฃ๐ ๐ฃ๐๐๐๐๐๐๐ 0 ๐๐ ๐๐๐ก ๐๐ ๐๐กโ ๐๐๐ฃ๐๐ ๐๐ ๐๐ข๐๐๐๐ก๐๐ก๐๐ฃ๐ ๐ฃ๐๐๐๐๐๐๐ i=1, 2, โฏ, ๐โ1 This can be done automatically in PROC GLM by using the CLASSS statement as we did in ANOVA. Any dummy variables defined with respect to a qualitative variable must be treated as a group. Individual t-tests are not meaningful. Partial F-tests must be performed on the group of dummy variables.
33
6-4 Other Aspects of Regression
Example 6-11 OPTIONS NOOVP NODATE NONUMBER; DATA EX611; INPUT FORM SCENT COLOR RESIDUE REGION QUALITY IF REGION=1 THEN REGION1=0; ELSE REGION1=1; /* IF REGION=1์ด๋ฉด REGION1=0 ์ด๊ณ THEN EAST IF REGION=2์ด๋ฉด REGION1=1 ์ด๊ณ THEN WEST */ FR=FORM*REGION1; RR=RESIDUE*REGION1; CARDS; PROC REG DATA=EX611; MODEL QUALITY=FORM RESIDUE REGION1; TITLE 'MODEL WITH DUMMY VARIABLE'; MODEL QUALITY=FORM RESIDUE REGION1 FR RR; TITLE 'INTERACTION MODEL WITH DUMMY VARIABLE'; RUN; QUIT;
34
6-4 Other Aspects of Regression
35
6-4 Other Aspects of Regression
36
6-4 Other Aspects of Regression
37
6-4 Other Aspects of Regression
38
6-4 Other Aspects of Regression
39
6-4 Other Aspects of Regression
40
6-3 Multiple Regression Example OPTIONS NOOVP NODATE NONUMBER;
DATA appraise; INPUT price units age size parking area cond$ IF COND='Fโ THEN COND1=1; ELSE COND1=0; IF COND='Gโ THEN COND2=1; ELSE COND2=0; CARDS; F G G E G G G G G G G F E G G G F E G F G E F E ods graphics on; PROC REG DATA=APPRAISE; MODEL PRICE=UNITS AGE AREA COND1 COND2/R; TITLE โREDUCED MODEL WITH DUMMY VARIABLE'; RUN; ods graphics off; QUIT;
41
6-3 Multiple Regression ๐น ๐ ๐ด๐บ๐ฌ Full Model 0.9801 0.9746 34123
Reduced Model 0.9771 0.9737 34721 With dummy 0.9860 0.9821 28673
42
6-3 Multiple Regression Example
43
6-3 Multiple Regression
44
6-3 Multiple Regression
45
6-3 Multiple Regression
46
Analysis of Covariance (๊ณต๋ถ์ฐ๋ถ์)
6-4 Other Aspects of Regression Analysis of Covariance (๊ณต๋ถ์ฐ๋ถ์) Suppose we have the following setup. 2 3 Treatment 1 2 r X Y โฏ ๐ 1,1 ๐ 1,1 ๐ 2,1 ๐ 2,1 ๐ ๐,1 ๐ ๐,1 โฎ ๐ 1,๐ ๐ 1,๐ ๐ 2,๐ ๐ 2,๐ ๐ ๐,๐ ๐ ๐,๐ 2 2 3 3 2 4 1 3 Y 2 1 3 2 4 1 3 4 1 2 4 1 3 3 1 4 4 4 1 4 1 X Suppose X and Y are linearly related. We are interested in comparing the means of Y at the different levels of the treatment. Suppose a plot of the data looks like the following. ๊ณต๋ถ์ฐ๋ถ์ (Analysis of Covariance : ANCOVA) ์ด๋ ๋ถ์ฐ๋ถ์๊ณผ ํ๊ท๋ถ์์ ํผํฉํ ํํ์ ๋ถ์์ผ๋ก์, ๋ถ์ฐ๋ถ์์์์ ๊ฐ์ด ๋
๋ฆฝ์ ์ธ r๊ฐ ๋ชจ์ง๋จ ํ๊ท (์ฒ๋ฆฌํจ๊ณผ)๋ค ๊ฐ์ ์ฐจ์ด๊ฐ ์๋๊ฐ๋ฅผ ๊ฒ์ฆํ๊ณ ์ ํ๋ ๊ฒฝ์ฐ์ ์ฌ์ฉ๋๋, ์๋ฃ (Y )๊ฐ ๋ค๋ฅธ ์ด๋ค ๋ณ์์ ํจ์๊ด๊ณ(์ฌ๊ธฐ์๋ 1์ฐจ์์ ๊ด๊ณ๋ง ๋ค๋ฃธ)์ ์๋ค๊ณ ๋ฏฟ์ด์ง ๋ ๋งค์ฐ ์ ์ฉํ ๋ถ์์๋จ์ด๋ค.
47
6-4 Other Aspects of Regression
Why Use Covariates? Concomitant variables or covariates are used to adjust for factors that influence the Y measurements. In randomized block designs, we did the same thing, but there we could control the value of the block variable. Now we assume we can measure the variable, but not control it. The plot on the previous page demonstrates why we need covariates in some situations. If the covariate (X) was ignored we would most likely conclude that treatment level 3 resulted in a larger mean than 1 and 4 but not different from 2. If the linear relation is extended we see that the value of Y in level 3 could very well be less than that of 1, nearly equal to that of 4 and surely less than that of 2. One assumption we need, equivalent to the no interaction assumption in two-way ANOVA, is that the slopes of the linear relationship between X and Y is the same in each treatment level.
48
Checking for Equal Slopes
6-4 Other Aspects of Regression Checking for Equal Slopes The Model we fit first Treatment = 1 ๐ 1๐ = ๐ฝ 0 + ๐ผ 1 +๐ฝ ๐ ๐ +๐ผ ๐ฝ 1 ๐ ๐ + ๐ 1๐ Y-intercept = ๐ฝ 0 + ๐ผ slope=๐ฝ+๐ผ ๐ฝ 1 โฎ Treatment = rโ1 ๐ (๐โ1)๐ = ๐ฝ 0 + ๐ผ ๐โ1 +๐ฝ ๐ ๐ +๐ผ ๐ฝ ๐โ1 ๐ ๐ + ๐ (๐โ1)๐ Y-intercept = ๐ฝ 0 + ๐ผ ๐โ slope=๐ฝ+๐ผ ๐ฝ ๐โ1 Treatment = r ๐ ๐๐ = ๐ฝ 0 +๐ฝ ๐ ๐ + ๐ ๐๐ Y-intercept = ๐ฝ slope=๐ฝ The test of equal slopes is ๐ป 0 : ๐ผ ๐ฝ 1 =๐ผ ๐ฝ 2 =โฏ=๐ผ ๐ฝ ๐โ1 =0 ๐ป 1 :๐ด๐ก ๐๐๐๐ ๐ก ๐๐๐ ๐๐๐ก ๐ง๐๐๐ If we fail to reject this we return the model without the interaction term and test without the interaction term and test ๐ป 0 : ๐ผ 1 = ๐ผ 2 =โฏ= ๐ผ ๐โ1 =0 ๐ป 1 :๐๐๐ก ๐๐๐ ๐ง๐๐๐
49
6-4 Other Aspects of Regression
EXAMPLE Four different formulations of an industrial glue are being tested. The tensile strength of the glue is also related to the thickness. Five observations on strength (Y) and thickness (X) in 0.01 inches are obtained for each formulation. The data are shown in the following table. Glue Formation 1 2 3 4 y x 46.5 45.9 49.8 46.1 44.3 13 14 12 48.7 49.0 50.1 48.5 45.2 10 11 46.3 47.1 48.9 48.2 50.3 15 44.7 43.0 51.0 48.1 48.6 16
50
6-4 Other Aspects of Regression
Example OPTIONS NOOVP NODATE NONUMBER; DATA GLUE; INPUT FORMULA STRENGTH THICK CARDS; Ods graphics on; PROC GLM DATA=GLUE; CLASS FORMULA; MODEL STRENGTH=FORMULA THICK FORMULA*THICK; TITLE 'ANALYSIS OF COVARIANCE WITH INTERACTION'; /* TEST FOR LINEARITY */ MODEL STRENGTH=FORMULA THICK/SOLUTION; LSMEANS FORMULA/PDIFF STDERR; TITLE 'ANALYSIS OF COVARIANCE WITHOUT INTERACTION'; /* TEST FOR Parallelism */ RUN; QUIT; SOLUTION produces a solution to the normal equations (parameter estimates). PROC GLM displays a solution by default when your model involves no classification variables, so you need this option only if you want to see the solution for models with classification effects. PDIFF requests that p-values for differences of the LS-means be produced. STDERR produces the standard error of the LS-means and the probability level for the hypothesisย H0: LS-mean=0
51
6-4 Other Aspects of Regression
PROC GLM DATA=GLUE; CLASS FORMULA; MODEL STRENGTH=FORMULA THICK FORMULA*THICK; TITLE 'ANALYSIS OF COVARIANCE WITH INTERACTION'; /* TEST FOR LINEARITYโ */ 4๊ฐ์ formula๊ฐ interaction์ด ์์ด parallelํ๊ฐ?
52
6-4 Other Aspects of Regression
53
6-4 Other Aspects of Regression
PROC GLM DATA=GLUE; CLASS FORMULA; MODEL STRENGTH=FORMULA THICK/SOLUTION; LSMEANS FORMULA/PDIFF STDERR; TITLE 'ANALYSIS OF COVARIANCE WITHOUT INTERACTION'; /* test for parallelism */ ์์์ ๊ฒฐ์ ํ parallel line์ด ๋์ผํ intercept๋ฅผ ์ง๋๋๊ฐ? ์์์ ๊ฒฐ์ ํ parallel line์ ๊ธฐ์ธ๊ธฐ๊ฐ zero์ธ๊ฐ?
54
6-4 Other Aspects of Regression
55
6-4 Other Aspects of Regression
56
6-4 Other Aspects of Regression
57
6-4 Other Aspects of Regression
58
6-4 Other Aspects of Regression
6-4.3 Variable Selection Procedures Best Subsets Regressions Selection Techniques R2 MSE Cp ๐ถ ๐ = ๐๐๐ธ(๐) ๐ 2 (๐น๐) โ๐+2๐
59
6-4 Other Aspects of Regression
6-4.3 Variable Selection Procedures Backward Elimination all regressors in the model t-test: smallest absolute t-value eliminated first Minitab ๐ผ=0.10 for cut-off form, residue, region
60
6-4 Other Aspects of Regression
6-4.3 Variable Selection Procedures Forward Selection No regressors in the model largest absolute t-value added first Minitab ๐ผ=0.25 for cut-off form, residue, region, scent
61
6-4 Other Aspects of Regression
6-4.3 Variable Selection Procedures Stepwise Regression begins with forward step, then backward elimination tin=tout Minitab ๐ผ=0.15 for cut-off form, residue, region
62
6-4 Other Aspects of Regression
Example OPTIONS NODATE NOOVP NONUMBER; DATA SALES; INFILE 'C:\Users\korea\Desktop\Working Folder 2017\imen214-stats\ch06\data\sales.dat'; INPUT SALES TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING; PROC CORR DATA=SALES; VAR SALES; WITH TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING; TITLE 'CORRELATIONS OF DEPENDENT WITH INDENDENTS'; VAR TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING; TITLE 'CORRELATIONS BETWEEN INDEPENDENT VARIABLES'; PROC REG DATA=SALES; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING/VIF R; TITLE 'REGRESSION MODEL WITH ALL VARIABLES'; PROC RSQUARE DATA=SALES CP; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING/ADJRSQ RMSE SSE SELECT=10; TITLE 'ALL POSSIBLE REGRESSIONS'; PROC STEPWISE DATA=SALES; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING/FORWARD; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING/BACKWARD; TITLE 'STEPWISE REGRESSION USING BACKWARD ELIMINATION'; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE ACCOUNTS WORKLOAD RATING; TITLE 'STEPWISE REGRESSION THE STEPWISE TECHNIQUE'; MODEL SALES=POTENT ADVERT SHARE ACCOUNTS/R; MODEL SALES=POTENT ADVERT SHARE CHANGE ACCOUNTS/R; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE/R; MODEL SALES=TIME POTENT ADVERT SHARE ACCOUNTS/R; MODEL SALES=TIME POTENT ADVERT SHARE CHANGE WORKLOAD/R; RUN; QUIT;
63
6-4 Other Aspects of Regression
64
6-4 Other Aspects of Regression
Example
65
6-4 Other Aspects of Regression
66
6-4 Other Aspects of Regression
67
6-4 Other Aspects of Regression
68
6-4 Other Aspects of Regression
69
6-4 Other Aspects of Regression
70
6-4 Other Aspects of Regression
71
All Possible Regressions
6-4 Other Aspects of Regression All Possible Regressions This is the brute force method of modeling. It is feasible if the number of independent variables is small (less than 10 or so) and the sample size is not too large. Some of the common quantities to look at are R-square should be large. Should be adequately increase when an additional variable is added. Adj R-square should not be much less than R-square. It should show an increase if a variable is added. Mallows Cp should be approximately the number of parameters in the model (including the y-intercept). This is a good measure to use to narrow down the possible models quickly, then use 1) and 2) to pick the final models. The model should make sense. Note: Many of the better methods of model selection are to time consuming to use on all possible regressions. A number of good models can be chosen and then use better methods.
72
6-4 Other Aspects of Regression
73
6-4 Other Aspects of Regression
74
6-4 Other Aspects of Regression
75
6-4 Other Aspects of Regression
76
6-4 Other Aspects of Regression
Stepwise Regression Forward Selection: Begins with no variables in the model. Calculates simple linear model for each X and adds most significant. (if above stated p-value). Calculates all models with already added variables and each non-added variable. Most significant is added. (if above sated p-value) This process is continued until no variables can be added. Backward Elimination: Model with all variables is fit. Least significant variable is removed (if p-value is greater than specified limit) and the model is refit without this variable. This process is continued until no variables can be removed.
77
6-4 Other Aspects of Regression
Stepwise Regression Stepwise Technique: This technique is a variation on the forward selection technique. After a variable is added, the least significant is also removed if it has a p-value greater than the specified limit. This accounts for multicollinearity to some degree. Typically you do not do a stepwise procedure if you do an all possible regressions and vice versa. Stepwise procedures are more economical than all possible regressions in large data sets. There is no guarantee that the stepwise procedures will end up with the same model or the โbestโ model.
78
6-4 Other Aspects of Regression
79
6-4 Other Aspects of Regression
80
6-4 Other Aspects of Regression
81
6-4 Other Aspects of Regression
82
6-4 Other Aspects of Regression
83
6-4 Other Aspects of Regression
84
6-4 Other Aspects of Regression
85
6-4 Other Aspects of Regression
86
6-4 Other Aspects of Regression
87
6-4 Other Aspects of Regression
88
6-4 Other Aspects of Regression
89
6-4 Other Aspects of Regression
90
6-4 Other Aspects of Regression
91
6-4 Other Aspects of Regression
92
6-4 Other Aspects of Regression
93
6-4 Other Aspects of Regression
94
6-4 Other Aspects of Regression
95
6-4 Other Aspects of Regression
96
6-4 Other Aspects of Regression
97
6-4 Other Aspects of Regression
98
6-4 Other Aspects of Regression
Press Statistic The main purpose of many regression analyses is to predict Y for a future set of Xโs. The problem is that we have only present Yโs and Xโs to use to make a model, but we would like to evaluate the model by how well it estimates Yโs with new Xโs. The Press Statistic tries to overcome this problem. It is similar to the DFFITS in that you remove one observation at a time. The parameters are then calculated and ๐ is calculated for the Xโs of the observation that is removed. Once the ๐ ๐ โs are calculated in this manner for each observation (call them ๐ ๐ โ ) the press statistic can be calculated. ๐๐๐๐ ๐ = ๐=1 ๐ ๐ ๐ โ ๐ ๐ โ 2 Notice that this is very similar to SSE. It is very computation intensive, however. The Press Statistic is obtained in SAS by using the r option on the model statement.
99
6-4 Other Aspects of Regression
Validation Data Split Split data into a fitting portion and a validation portion. This should be done randomly. Perform the model fitting routine as discussed earlier using data in the fitting portion only. For each viable model compute the SSE using the observations in the validation data portion. The best model is the one that minimizes the SSE. Recalculate the chosen model using the entire data set. Notice this procedure requires a large enough data set to enable you to split a validation portion off and still have adequate data to evaluate models. The process is tedious in SAS, requiring multiple runs or fancy programming.
100
6-4 Other Aspects of Regression
101
6-4 Other Aspects of Regression
102
6-4 Other Aspects of Regression
103
6-4 Other Aspects of Regression
104
6-4 Other Aspects of Regression
105
6-4 Other Aspects of Regression
106
6-4 Other Aspects of Regression
107
6-4 Other Aspects of Regression
108
6-4 Other Aspects of Regression
109
6-4 Other Aspects of Regression
110
6-4 Other Aspects of Regression
111
6-4 Other Aspects of Regression
112
6-4 Other Aspects of Regression
113
6-4 Other Aspects of Regression
114
6-4 Other Aspects of Regression
115
6-4 Other Aspects of Regression
116
6-4 Other Aspects of Regression
117
6-4 Other Aspects of Regression
118
6-4 Other Aspects of Regression
119
6-4 Other Aspects of Regression
120
6-4 Other Aspects of Regression
121
6-4 Other Aspects of Regression
122
6-4 Other Aspects of Regression
123
6-4 Other Aspects of Regression
124
6-4 Other Aspects of Regression
125
6-4 Other Aspects of Regression
Model R2 Adj R2 MSE PRESS Potent advert share accounts 0.9004 0.8805 Potent advert share change accounts 0.9119 0.8888 Time potent advert share change 0.9108 0.8873 Time potent advert share accounts 0.9064 0.8817 Time potent advert share change workload 0.9109 0.8812 No one model could be used ๏ confidence interval might be helpful to decide the best model or parsimony
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.