Presentation is loading. Please wait.

Presentation is loading. Please wait.

Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park.

Similar presentations


Presentation on theme: "Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park."— Presentation transcript:

1 Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park

2 In many problems one or more variables are related, it is of interest to model and explore this relationship. The model can be used for prediction, process optimization, or process control.

3 Modelling

4 Interpolation Method interpolation is a method of constructing new data points within the range of a discrete set of known data points.discrete set In engineering and science one often has a number of data points, as obtained by sampling or experiment, and tries to construct a function which closely fits those data points.engineeringsciencesamplingexperiment This is called curve fitting or regression analysis. Interpolation is a specific case of curve fitting, in which the function must go exactly through the data points.curve fittingregression analysis

5 xf(x) 00 10.8415 20.9093 2.5? 30.1411 4-0.7568 5-0.9589 6-0.2794

6 Piecewise constant interpolation The simplest interpolation method is to locate the nearest data value, and assign the same value. In one dimension, there are seldom good reasons to choose this one over linear interpolation, which is almost as cheap, but in higher dimensions, in multivariate interpolation, this can be a favourable choice for its speed and simplicitymultivariate interpolation

7 Linear interpolation One of the simplest methods is linear interpolation (sometimes known as lerp). Consider the above example of determining f(2.5). Since 2.5 is midway between 2 and 3, it is reasonable to take f(2.5) midway between f(2) = 0.9093 and f(3) = 0.1411, which yields 0.5252.linear Generally, linear interpolation takes two data points, say (x a,y a ) and (x b,y b ). Linear interpolation is quick and easy, but it is not very precise. Another disadvantage is that the interpolant is not differentiable at the point x k.differentiable

8 Polynomial interpolation Polynomial interpolation is a generalization of linear interpolation. Note that the linear interpolant is a linear function. We now replace this interpolant by a polynomial of higher degree.linear function polynomialdegree Consider again the problem given above. The following sixth degree polynomial goes through all the seven points: f(x) = − 0.0001521x6 − 0.003130x5 + 0.07321x4 − 0.3577x3 + 0.2255x2 + 0.9038x. Substituting x = 2.5, we find that f(2.5) = 0.5965.

9 Spline interpolation Remember that linear interpolation uses a linear function for each of intervals [xk,xk+1]. Spline interpolation uses low-degree polynomials in each of the intervals, and chooses the polynomial pieces such that they fit smoothly together. The resulting function is called a spline.spline

10 Extrapolation Method A sound choice of which extrapolation method to apply relies on a prior knowledge of the process that created the existing data points. Crucial questions are for example if the data can be assumed to be continuous, smooth, possibly periodic etc.

11 Linear extrapolation Polynomial extrapolation Conic extrapolation A conic section can be created using five points near the end of the known data. If the conic section created is an ellipse or circle, it will loop back and rejoin itself. A parabolic or hyperbolic curve will not rejoin itself, but may curve back relative to the X-axis. This type of extrapolation could be done with a conic sections template (on paper) or with a computer.

12 Simple Linear Regression and Correlation (Empirical Models) obseravtion number Hydrocarbon level x(%) Purity y(%) 10.9990.01 21.0289.05 31.1591.43 41.2993.74 51.4696.73 61.3694.45 70.8787.59 81.2391.77 91.5599.42 101.493.65 111.1993.54 121.1592.52 130.9890.56 141.0189.54 151.1189.85 161.290.39 171.2693.25 181.3293.41 191.4394.98 200.9587.33 y : the purity of oxygen produced in a chemical distillation process x : the percentage of hydrocarbons that are present in the main condenser of the distillation unit Example 6-1

13 Simple Linear Regression and Correlation Stat > Regression > Fitted Line Plot

14 Simple Linear Regression and Correlation (Empirical Models) Example 6-1

15 Regression Analysis: Purity y(%) versus Hydrocarbon level x(%) The regression equation is Purity y(%) = 74.28 + 14.95 Hydrocarbon level x(%) S = 1.08653 R-Sq = 87.7% R-Sq(adj) = 87.1% Analysis of Variance Source DF SS MS F P Regression 1 152.127 152.127 128.86 0.000 Error 18 21.250 1.181 Total 19 173.377 Simple Linear Regression and Correlation (Empirical Models)

16 Example 6-1 Simple Linear Regression and Correlation (Empirical Models)

17 Example 6-1 Simple Linear Regression and Correlation (Empirical Models) Predicted Values for New Observations New Obs Fit SE Fit 95% CI 95% PI 1 89.081 0.364 (88.316, 89.846) (86.674, 91.489) 2 89.530 0.336 (88.824, 90.235) (87.141, 91.919) 3 91.473 0.250 (90.947, 91.999) (89.130, 93.815) 4 93.566 0.273 (92.993, 94.138) (91.212, 95.919) 5 96.107 0.424 (95.216, 96.998) (93.656, 98.557) 6 94.612 0.325 (93.929, 95.295) (92.229, 96.995) 7 87.288 0.493 (86.251, 88.324) (84.781, 89.795) 8 92.669 0.247 (92.150, 93.188) (90.328, 95.010) 9 97.452 0.526 (96.348, 98.556) (94.916, 99.988) 10 95.210 0.362 (94.449, 95.971) (92.804, 97.616) 11 92.071 0.243 (91.560, 92.582) (89.732, 94.410) 12 91.473 0.250 (90.947, 91.999) (89.130, 93.815) 13 88.932 0.374 (88.146, 89.718) (86.518, 91.346) 14 89.380 0.345 (88.655, 90.105) (86.985, 91.775) 15 90.875 0.268 (90.312, 91.438) (88.524, 93.226) 16 92.220 0.243 (91.710, 92.731) (89.881, 94.559) 17 93.117 0.257 (92.577, 93.657) (90.771, 95.463) 18 94.014 0.293 (93.399, 94.629) (91.650, 96.378) 19 95.658 0.392 (94.834, 96.483) (93.231, 98.085) 20 88.483 0.405 (87.633, 89.334) (86.047, 90.919)

18 Example 6-1 Simple Linear Regression and Correlation (Empirical Models) Hydrocarbon Purity Obs level x(%) y(%) Fit SE Fit Residual St Resid 1 0.99 90.010 89.081 0.364 0.929 0.91 2 1.02 89.050 89.530 0.336 -0.480 -0.46 3 1.15 91.430 91.473 0.250 -0.043 -0.04 4 1.29 93.740 93.566 0.273 0.174 0.17 5 1.46 96.730 96.107 0.424 0.623 0.62 6 1.36 94.450 94.612 0.325 -0.162 -0.16 7 0.87 87.590 87.288 0.493 0.302 0.31 8 1.23 91.770 92.669 0.247 -0.899 -0.85 9 1.55 99.420 97.452 0.526 1.968 2.07R 10 1.40 93.650 95.210 0.362 -1.560 -1.52 11 1.19 93.540 92.071 0.243 1.469 1.39 12 1.15 92.520 91.473 0.250 1.047 0.99 13 0.98 90.560 88.932 0.374 1.628 1.60 14 1.01 89.540 89.380 0.345 0.160 0.16 15 1.11 89.850 90.875 0.268 -1.025 -0.97 16 1.20 90.390 92.220 0.243 -1.830 -1.73 17 1.26 93.250 93.117 0.257 0.133 0.13 18 1.32 93.410 94.014 0.293 -0.604 -0.58 19 1.43 94.980 95.658 0.392 -0.678 -0.67 20 0.95 87.330 88.483 0.405 -1.153 -1.14

19 Example 6-1 Simple Linear Regression and Correlation (Empirical Models)

20 Example 6-1 Simple Linear Regression and Correlation (Empirical Models)

21 Multiple Linear Regression Model observatio nTempRateViscosity 18082256 29392340 3100102426 482122293 590112330 69982368 78182250 896102409 994122364 1093112379 1197132440 1295112364 1310082404 1485122317 158692309 1687122328

22 Multiple Linear Regression Model Stat > Regression > Regression

23 Multiple Linear Regression Model Regression Analysis: Viscosity versus Temp, Rate The regression equation is Viscosity = 1566 + 7.62 Temp + 8.58 Rate Predictor Coef SE Coef T P Constant 1566.08 61.59 25.43 0.000 Temp 7.6213 0.6184 12.32 0.000 Rate 8.585 2.439 3.52 0.004 S = 16.3586 R-Sq = 92.7% R-Sq(adj) = 91.6% Analysis of Variance Source DF SS MS F P Regression 2 44157 22079 82.50 0.000 Residual Error 13 3479 268 Total 15 47636

24 Multiple Linear Regression Model

25


Download ppt "Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park."

Similar presentations


Ads by Google