Download presentation
Presentation is loading. Please wait.
1
Engineering Computation
Part 3 E. T. S. I. Caminos, Canales y Puertos
2
Learning Objectives for Lecture
1. Motivate Study of Systems of Equations and particularly Systems of Linear Equations 2. Review steps of Gaussian Elimination 3. Examine how roundoff error can enter and be magnified in Gaussian Elimination 4. Introduce Pivoting and Scaling as defenses against roundoff. 5. Consider what an engineer can do to generate well formulated problems. E. T. S. I. Caminos, Canales y Puertos
3
Systems of Equations In Part 2 we have tried to determine the value x, satisfying f(x)=0. In this part we try to obtain the values x1,x2, xn, satisfying the system of equations: These systems can be linear or nonlinear, but in this part we deal with linear systems: E. T. S. I. Caminos, Canales y Puertos
4
Systems of Equations where a and b are constant coefficients, and n is the number of equations. Many of the engineering fundamental equations are based on conservation laws. In mathematical terms, these principles lead to balance or continuity equations relating the system behavior with respect to the amount of the magnitude being modelled and the extrenal stimuli acting on the system. E. T. S. I. Caminos, Canales y Puertos
5
Systems of Equations Column 3 Row 2
Matrices are rectangular sets of elements represented by a single symbol. If the set if horizontal it is called row, and if it is vertical, it is called column. Column 3 Row 2 Column vector Row vector E. T. S. I. Caminos, Canales y Puertos
6
Systems of Equations There are some special types of matrices:
Identity matrix Symmetric matrix Diagonal matrix Upper triangular matrix E. T. S. I. Caminos, Canales y Puertos
7
Lower triangular matrix
Systems of Equations Lower triangular matrix Banded matrix Half band width All elements are null with the exception of thoise in a band centered around the main diagonal. This matrix has a band width of 3 and has the name of tridiagonal. E. T. S. I. Caminos, Canales y Puertos
8
Systems of Equations Linear Algebraic Equations
a11x1 + a12x2 + a13x3 + … + a1nxn = b1 a21x1 + a22x2 + a23x3 + … + a2nxn = b2 ….. an1x1 + an2x2 + an3x3 + … + anxn = bn where all aij's and bi's are constants. In matrix form: n x n n x 1 n x 1 or simply [A]{x} = {b} E. T. S. I. Caminos, Canales y Puertos
9
Systems of Equations Matrix product: Matrix representation of a system
Resulting dimensions E. T. S. I. Caminos, Canales y Puertos
10
Systems of Equations Graphic Solution: Systems of equations are hyperplanes (straight lines, planes, etc.). The solution of a system is the intersection of these hyperplanes. Compatible and determined system. Vectors are linearly independent. Unique solution. Determinant of A is non-null. E. T. S. I. Caminos, Canales y Puertos
11
Systems of Equations Incompatible system, Linearly dependent vectors. Null determinant of A. There is no solution. Compatible but undetermined system. Linearly dependent vectors. Null determinant of A. There exists an infinite number of solutions. E. T. S. I. Caminos, Canales y Puertos
12
Systems of Equations Compatible and determined system. Linearly independent vectors. Nonnull determinant of A, but close to zero. There exists a solution but it is difficult to find precisely. It is an ill conditioned system leading to numerical errors. E. T. S. I. Caminos, Canales y Puertos
13
Gauss elimination Naive Gauss elimination method: The Gauss’ method has two phases: Forward elimination and backsustitution. In the first, the system is reduced to an upper triangular system: First, the unknown x1 is eliminated. To this end, the first row is multiplied by -a21/a11 and added to the second row. The same is done with all other succesive rows (n-1 times) until only the first equation contains the first unknown x1. Pivot equation substract pivot E. T. S. I. Caminos, Canales y Puertos
14
Gauss elimination This operation is repeated with all variables xi, until an upper triangular matrix is obtained. Next, the system is solved by backsustitution. The number of operations (FLOPS) used in the Gauss method is: Pass 1 Pass 2 E. T. S. I. Caminos, Canales y Puertos
15
Gauss elimination 1. Forward Elimination (Row Manipulation):
a. Form augmented matrix [A|b]: b. By elementary row manipulations, reduce [A|b] to [U|b'] where U is an upper triangular matrix: DO i = 1 to n-1 DO k = i+1 to n Row(k) = Row(k) - (aki/aii)*Row(i) ENDDO E. T. S. I. Caminos, Canales y Puertos
16
Gauss elimination 2. Back Substitution
Solve the upper triangular system [U]{x} = {b´} xn = b'n / unn DO i = n-1 to 1 by (-1) END E. T. S. I. Caminos, Canales y Puertos
17
Gauss elimination (example)
Consider the system of equations To 2 significant figures, the exact solution is: We will use 2 decimal digit arithmetic with rounding. E. T. S. I. Caminos, Canales y Puertos
18
Gauss elimination (example)
Start with the augmented matrix: Multiply the first row by –1/50 and add to second row. Multiply the first row by –2/50 and add to third row: Multiply the second row by –6/40 E. T. S. I. Caminos, Canales y Puertos
19
Gauss elimination (example)
Now backsolve: (vs , et = 2.2%) (vs , et = 2.5%) (vs , et = 0%) E. T. S. I. Caminos, Canales y Puertos
20
Gauss elimination (example)
Consider an alternative solution interchanging rows: After forward elimination, we obtain: Now backsolve: x3 = (vs , et = 4.4%) x2 = (vs , et = 50%) x1 = (vs , et = 100%) Apparently, the order of the equations matters! E. T. S. I. Caminos, Canales y Puertos
21
Gauss elimination (example)
WHAT HAPPENED? When we used 50 x1 + 1 x2 + 2 x3 = 1 to solve for x1, there was little change in other equations. When we used 2 x1 + 6 x x3 = 3 to solve for x1 it made BIG changes in the other equations. Some coefficients for other equations were lost! The second equation has little to do with x1. It has mainly to do with x3. As a result we obtained LARGE numbers in the table, significant roundoff error occurred and information was lost. Things didn't go well! If scaling factors | aji / aii | are 1 then the effect of roundoff errors is diminished. E. T. S. I. Caminos, Canales y Puertos
22
Gauss elimination (example)
Effect of diagonal dominance: As a first approximation roots are: xi bi / aii Consider the previous examples: E. T. S. I. Caminos, Canales y Puertos
23
Gauss elimination (example)
Goals: 1. Best accuracy (i.e. minimize error) 2. Parsimony (i.e. minimize effort) Possible Problems: A. Zero on diagonal term ÷ by zero. B. Many floating point operations (flops) cause numerical precision problems and propagation of errors. C. System may be ill-conditioned: det[A] 0. D. No solution or an infinite # of solutions: det[A] = 0. Possible Remedies: A. Carry more significant figures (double precision). B. Pivot when the diagonal is close to zero. C. Scale to reduce round-off error. E. T. S. I. Caminos, Canales y Puertos
24
Gauss elimination (pivoting)
A. Row pivoting (Partial Pivoting) - In any good routine, at each step i, find maxk | aki | for k = i, i+1, i+2, ..., n Move corresponding row to pivot position. (i) Avoids zero aii (ii) Keeps numbers small & minimizes round-off, (iii) Uses an equation with large | aki | to find xi Maintains diagonal dominance. Row pivoting does not affect the order of the variables. Included in any good Gaussian Elimination routine. E. T. S. I. Caminos, Canales y Puertos
25
Gauss elimination (pivoting)
B. Column pivoting - Reorder remaining variables xj for j = i, ,n so get largest | aji | Column pivoting changes the order of the unknowns, xi, and thus leads to complexity in the algorithm. Not usually done. C. Complete or Full pivoting Performing both row pivoting and column pivoting. (If [A] is symmetric, needed to preserve symmetry.) E. T. S. I. Caminos, Canales y Puertos
26
Gauss elimination (pivoting)
How to fool pivoting: Multiply the third equation by 100 and then performing pivoting will yield: Forward elimination then yields (2-digit arithmetic): Backsolution yields: x3 = (vs , et = 4.4%) x2 = (vs , et = 50.0%) x1 = (vs , et = 100%) The order of the rows is still poor!! E. T. S. I. Caminos, Canales y Puertos
27
Gauss elimination (scaling)
A. Express all equations (and variables) in comparable units so all elements of [A] are about the same size. B. If that fails, and maxj |aij| varies widely across the rows, replace each row i by: aij This makes the largest coefficient |aij| of each equation equal to 1 and the largest element of [A] equal to 1 or -1 NOTE: Routines generally do not scale automatically; scaling can cause round-off error too! SOLUTIONS • Don't actually scale, but use hypothetical scaling factors to determine what pivoting is necessary. • Scale only by powers of 2: no roundoff or division required. E. T. S. I. Caminos, Canales y Puertos
28
Gauss elimination (scaling)
How to fool scaling: A poor choice of units can undermine the value of scaling. Begin with our original example: If the units of x1 were expressed in µg instead of mg the matrix might read: Scaling then yields: Which equation is used to determine x1 ? Why bother to scale ? E. T. S. I. Caminos, Canales y Puertos
29
Gauss elimination (operation counting)
In numerical scientific calculations, the number of multiplies & divides often determines CPU time. (This represents the numerical effort!) One floating point multiply or divide (plus any associated adds or subtracts) is called a FLOP. (The adds/subtracts use little time compared to the multiplies/divides.) FLOP = FLoating point OPeration. Examples: a * x + b a / x – b E. T. S. I. Caminos, Canales y Puertos
30
Gauss elimination (operation counting)
Useful identities in counting FLOPS: O(mn) means that there are terms of order mn and lower. E. T. S. I. Caminos, Canales y Puertos
31
Gauss elimination (operation counting)
Simple Example of Operation Counting: DO i = 1 to n Y(i) = X(i)/i – 1 ENDDO X(i) and Y(i) are arrays whose values change when i changes. In each iteration X(i)/i – 1 represents one FLOP because it requires one division (& one subtraction). The DO loop extends over i from 1 to n iterations: E. T. S. I. Caminos, Canales y Puertos
32
Gauss elimination (operation counting)
Another Example of Operation Counting: DO i = 1 to n Y(i) = X(i) X(i) + 1 DO j = i to n Z(j) = [ Y(j) / X(i) ] Y(j) + X(i) ENDDO With nested loops, always start from the innermost loop. [Y(j)/X(i)] * Y(j) + X(i) represents 2 FLOPS E. T. S. I. Caminos, Canales y Puertos
33
Gauss elimination (operation counting)
For the outer i-loop: X(i) • X(i) + 1 represents 1 FLOP = 3n +2n2 - n2 - n = n2 + 2n = n2 + O(n) E. T. S. I. Caminos, Canales y Puertos
34
Gauss elimination (operation counting)
Forward Elimination: DO k = 1 to n–1 DO i = k+1 to n r = A(i,k)/A(k,k) DO j = k+1 to n A(i,j)=A(i,j) – r*A(k,j) ENDDO B(i) = B(i) – r*B(k) E. T. S. I. Caminos, Canales y Puertos
35
Gauss elimination (operation counting)
Operation Counting for Gaussian Elimination Back Substitution: X(n) = B(n)/A(n,n) DO i = n–1 to 1 by –1 SUM = 0 DO j = i+1 to n SUM = SUM + A(i,j)*X(j) ENDDO X(i) = [B(i) – SUM]/A(i,i) E. T. S. I. Caminos, Canales y Puertos
36
Gauss elimination (operation counting)
Operation Counting for Gaussian Elimination Forward Elimination Inner loop: Second loop: = (n2 + 2n) – 2(n + 1)k + k2 E. T. S. I. Caminos, Canales y Puertos
37
Gauss elimination (operation counting)
Operation Counting for Gaussian Elimination Forward Elimination (cont'd) Outer loop = E. T. S. I. Caminos, Canales y Puertos
38
Gauss elimination (operation counting)
Operation Counting for Gaussian Elimination Back Substitution Inner Loop: Outer Loop: E. T. S. I. Caminos, Canales y Puertos
39
Gauss elimination (operation counting)
Total flops = Forward Elimination + Back Substitution = n3/3 + O (n2) + n2/ O (n) n3/3 + O (n2) To convert (A,b) to (U,b') requires n3/3, plus terms of order n2 and smaller, flops. To back solve requires: n = n (n+1) / 2 flops; Grand Total: the entire effort requires n3/3 + O(n2) flops altogether. E. T. S. I. Caminos, Canales y Puertos
40
Gauss-Jordan Elimination
Diagonalization by both forward and backward elimination in each column. Perform elimination both backwards and forwards until: Operation count for Gauss-Jordan is: (slower than Gauss elimination) E. T. S. I. Caminos, Canales y Puertos
41
Gauss-Jordan Elimination
Example (two-digit arithmetic): x1 = (vs , et = 6.3%) x2 = (vs , et = 0%) x3 = (vs , et = 2.2%) E. T. S. I. Caminos, Canales y Puertos
42
Gauss-Jordan Matrix Inversion
The solution of: [A]{x} = {b} is: {x} = [A]-1{b} where [A]-1 is the inverse matrix of [A] Consider: [A] [A]-1 = [ I ] 1) Create the augmented matrix: [ A | I ] 2) Apply Gauss-Jordan elimination: ==> [ I | A-1 ] E. T. S. I. Caminos, Canales y Puertos
43
Gauss-Jordan Matrix Inversion
Gauss-Jordan Matrix Inversion (with 2 digit arithmetic): MATRIX INVERSE [A-1] E. T. S. I. Caminos, Canales y Puertos
44
Gauss-Jordan Matrix Inversion
CHECK: [ A ] [ A ]-1 = [ I ] [ A ]-1 { b } = { x } Gaussian Elimination E. T. S. I. Caminos, Canales y Puertos
45
LU decomposition LU decomposition - The LU decomposition is a method that uses the elimination techniques to transform the matrix A in a product of triangular matrices. This is specially useful to solve systems with different vectors b, because the same decomposition of matrix A can be used to evaluate in an efficient form, by forward and backward sustitution, all cases. E. T. S. I. Caminos, Canales y Puertos
46
LU decomposition Initial system Decomposition Transformed system 1
Substitution Transformed system 2 Backward sustitution Forward sustitution E. T. S. I. Caminos, Canales y Puertos
47
LU decomposition LU decomposition is very much related to Gauss method, because the upper triangular matrix is also looked for in the LU decomposition. Thus, only the lower triangular matrix is needed. Surprisingly, during the Gauss elimination procedure, this matrix L is obtained, but one is not aware of this fact. The factors we use to get zeroes below the main diagonal are the elements of this matrix L. Substract E. T. S. I. Caminos, Canales y Puertos
48
LU decomposition resto resto E. T. S. I. Caminos, Canales y Puertos
49
LU decomposition (Complexity)
Basic Approach Consider [A]{x} = {b} a) Gauss-type "decomposition" of [A] into [L][U] n3/3 flops [A]{x} = {b} becomes [L] [U]{x} = {b}; let [U]{x} {d} b) First solve [L] {d} = {b} for {d} by forward subst n2/2 flops c) Then solve [U]{x} = {d} for {x} by back substitution n2/2 flops E. T. S. I. Caminos, Canales y Puertos
50
LU Decompostion: notation
[A] = [L] + [U0] [A] = [L0] + [U] [A] = [L0] + [U0] + [D] [A] = [L1] [U] [A] = [L] [U1] E. T. S. I. Caminos, Canales y Puertos
51
LU decomposition LU Decomposition Variations
Doolittle [L1][U] General [A] Crout [L][U1] General [A] Cholesky [L][L] T Pos. Def. Symmetric [A] Cholesky works only for Positive Definite Symmetric matrices Doolittle versus Crout: • Doolittle just stores Gaussian elimination factors where Crout uses a different series of calculations (see C&C ). • Both decompose [A] into [L] and [U] in n3/3 FLOPS • Different location of diagonal of 1's • Crout uses each element of [A] only once so the same array can be used for [A] and [L\U] saving computer memory! E. T. S. I. Caminos, Canales y Puertos
52
Definition of a matrix inverse:
LU decomposition Matrix Inversion Definition of a matrix inverse: [A] [A]-1 = [ I ] ==> [A] {x} = {b} [A]-1 {b} = {x} First Rule: Don’t do it. (numerically unstable calculation) E. T. S. I. Caminos, Canales y Puertos
53
LU decomposition Matrix Inversion If you really must --
1) Gaussian elimination: [A | I ] –> [U | B'] ==> A-1 2) Gauss-Jordan: [A | I ] ==> [I | A-1 ] Inversion will take n3 + O(n2) flops if one is careful about where zeros are (taking advantage of the sparseness of the matrix) Naive applications (without optimization) take 4n3/3 + O(n2) flops. For example, LU decomposition requires n3/3 + O(n2) flops. Back solving twice with n unit vectors ei: 2 n (n2/2) = n3 flops. Altogether: n3/3 + n3 = 4n3/3 + O(n2) flops E. T. S. I. Caminos, Canales y Puertos
54
FLOP Counts for Linear Algebraic Equations
Summary FLOP Counts for Linear Algebraic Equations, [A]{x} = {b} Gaussian Elimination (1 r.h.s) n3/3 + O (n2) Gauss-Jordan (1 r.h.s) n3/2 + O (n2) LU decomposition n3/3 + O (n2) Each extra LU right-hand-side n2 Cholesky decomposition (symmetric A) n3/6 + O (n2) Inversion (naive Gauss-Jordan) 4n3/3 +O (n2) Inversion (optimal Gauss-Jordan) n3 + O (n2) Solution by Cramer's Rule n! E. T. S. I. Caminos, Canales y Puertos
55
Errors in Solutions to Systems of Linear Equations
System of Equations Errors in Solutions to Systems of Linear Equations Objective: Solve [A]{x} = {b} Problem: Round-off errors may accumulate and even be exaggerated by the solution procedure. Errors are often exaggerated if the system is ill-conditioned Possible remedies to minimize this effect: 1. Partial or complete pivoting 2. Work in double precision 3. Transform the problem into an equivalent system of linear equations by scaling or equilibrating E. T. S. I. Caminos, Canales y Puertos
56
Errors in Solutions to Systems of Linear Equations
Ill-conditioning A system of equations is singular if det|A| = 0 If a system of equations is nearly singular it is ill-conditioned. Systems which are ill-conditioned are extremely sensitive to small changes in coefficients of [A] and {b}. These systems are inherently sensitive to round-off errors. Question: Can we develop a means for detecting these situations? E. T. S. I. Caminos, Canales y Puertos
57
Errors in Solutions to Systems of Linear Equations
Ill-conditioning of [A]{x} = {b}: Consider the graphical interpretation for a 2-equation system: We can plot the two linear equations on a graph of x1 vs. x2. x1 x2 b2/a22 b1/a11 b1/a12 a11x1+ a12x2 = b1 a21x1+ a22x2 = b2 b2/a21 E. T. S. I. Caminos, Canales y Puertos
58
Errors in Solutions to Systems of Linear Equations
Ill-conditioning of [A]{x} = {b}: Consider the graphical interpretation for a 2-equation system: We can plot the two linear equations on a graph of x1 vs. x2. x1 x1 x2 x2 Uncertainty in x2 Uncertainty in x2 Well-conditioned Ill-conditioned E. T. S. I. Caminos, Canales y Puertos
59
Errors in Solutions to Systems of Linear Equations
Ways to detect ill-conditioning: 1. Calculate {x}, make small change in [A] or {b} and determine change in solution {x}. 2. After forward elimination, examine diagonal of upper triangular matrix. If aii << ajj, i.e. there is a relatively small value on diagonal, then this may indicate ill-conditioning. 3. Compare {x}single with {x}double 4. Estimate "condition number" for A. Substituting the calculated {x} into [A]{x} and checking this against {b} will not always work!!! E. T. S. I. Caminos, Canales y Puertos
60
Errors in Solutions to Systems of Linear Equations
Ways to detect ill-conditioning: If det|A| = 0 the matrix is singular ==> the determinant may be an indicator of conditioning If det|A| is near zero is the matrix ill-conditioned? Consider: After scaling: ==> det|A| will provide an estimate of conditioning if it is normalized by the "magnitude" of the matrix. E. T. S. I. Caminos, Canales y Puertos
61
Norms Norms and the Condition Number
We need a quantitative measure of ill-conditioning. This measure will then directly reflect the possible magnitude of round off effects. To do this we need to understand norms: Norm: Scalar measure of the magnitude of a matrix or vector ("how big" a vector is). Not to be confused with the dimension of a matrix. E. T. S. I. Caminos, Canales y Puertos
62
Vector Norms Vector Norms: Scalar measure of the magnitude of a vector
Here are some vector norms for n x 1 vectors {x} with typical elements xi. Each is in the general form of a p norm defined by the general relationship: 1. Sum of the magnitudes: 2. Magnitude of largest element: (infinity norm) 3. Length or Euclidean norm: E. T. S. I. Caminos, Canales y Puertos
63
Norms Vector Norms Required Properties of vector norm:
1. ||x|| 0 and ||x|| = 0 if and only if [x]=0 2 ||kx|| = k ||x|| where k is any positive scalar 3. ||x+y|| ||x|| + ||y|| Triangle Inequality For the Euclidean vector norm we also have 4. ||x•y|| ||x|| ||y|| because the dot product or inner product property satisfies: ||xy|| = ||x||•||y|| |cos()| ||x|| • ||y||. E. T. S. I. Caminos, Canales y Puertos
64
Matrix Norms Matrix Norms: Scalar measure of the magnitude of a matrix. Matrix norms corresponding to vector norms above are defined by the general relationship: 1. Largest column sum: (column sum norm) 2. Largest row sum: (row sum norm) (infinity norm) E. T. S. I. Caminos, Canales y Puertos
65
Matrix norms 3. Spectral norm: ||A|| 2 = (µmax)1/2
where µmax is the largest eigenvalue of [A]T[A] If [A] is symmetric, (µmax)1/2 = max , is the largest eigenvalue of [A]. (Note: this is not the same as the Euclidean or Frobenius norm, seldom used: E. T. S. I. Caminos, Canales y Puertos
66
Matrix norms Matrix Norms
For matrix norms to be useful we require that 0. || Ax || || A || ||x || General properties of any matrix norm: 1. || A || 0 and || A || = 0 iff [A] = 0 2. || k A || = k || A || where k is any positive scalar 3. || A + B || || A || + || B || "Triangle Inequality" 4. || A B || || A || || B || Why are norms important? Norms permit us to express the accuracy of the solution {x} in terms of || x || Norms allow us to bound the magnitude of the product [ A ] {x} and the associated errors. E. T. S. I. Caminos, Canales y Puertos
67
Error Analysis Forward and backward error analysis can estimate the effect of truncation and roundoff errors on the precision of a result. The two approaches are alternative views: Forward (a priori) error analysis tries to trace the accumulation of error through each process of the algorithm, comparing the calculated and exact values at every stage. Backward (a posteriori) error analysis views the final solution as the exact solution to a perturbed problem. One can consider how different the perturbed problem is from the original problem. Here we use the condition number of a matrix [A] to specify the amount by which relative errors in [A] and/or {b} due to input, truncation, and rounding can be amplified by the linear system in the computation of {x}. E. T. S. I. Caminos, Canales y Puertos
68
Error Analysis Backward Error Analysis of [A]{x} = {b} for errors in {b} Suppose the coefficients {b} are not precisely represented. What might be the effect on the calculated value for {x + dx}? Lemma: [A]{x} = {b} yields ||A|| ||x|| ||b|| or Now an error in {b} yields a corresponding error in {x}: [A ]{x + dx} = {b + db} [A]{x} + [A]{ dx} = {b} + {db} Subtracting [A]{x} = {b} yields: [A]{dx} = {db} ––> {dx} = [A]-1{db} E. T. S. I. Caminos, Canales y Puertos
69
Error Analysis Backward Error Analysis of [A]{x} = {b} for errors in {b} Taking norms we have: And using the lemma: we then have : Define the condition number as k = cond [A] ||A-1|| ||A|| 1 If k 1 or k is small, the system is well-conditioned If k >> 1, system is ill conditioned. 1 = || I || = || A-1A || || A-1 || || A || = k = Cond(A) E. T. S. I. Caminos, Canales y Puertos
70
Error Analysis Backward Error Analysis of [A]{x} = {b} for errors in [A] If the coefficients in [A] are not precisely represented, what might be effect on the calculated value of {x+ dx}? [A + dA ]{x + dx} = {b} [A]{x} + [A]{ dx} + [dA]{x+dx} = {b} Subtracting [A]{x} = {b} yields: [A]{ dx} = – [dA]{x+dx} or {dx} = – [A]-1 [dA] {x+dx} Taking norms and multiplying by || A || / || A || yields : E. T. S. I. Caminos, Canales y Puertos
71
E. T. S. I. Caminos, Canales y Puertos
Linfield & Penny, 1999
72
E. T. S. I. Caminos, Canales y Puertos
Linfield & Penny, 1999
73
Error Analysis Estimate of Loss of Significance:
Consider the possible impact of errors [dA] on the precision of {x}. implies that if Or, taking log of both sides: s > p - log10() log10() is the loss in decimal precision; i.e., we start with p decimal figures and end-up with s decimal figures. It is not always necessary to find [A]-1 to estimate k = cond[A]. Instead, use an estimate based upon iteration of inverse matrix using LU decomposition. E. T. S. I. Caminos, Canales y Puertos
74
Iterative Solution Methods
Impetus for Iterative Schemes: 1. May be more rapid if coefficient matrix is "sparse" 2. May be more economical with respect to memory 3. May also be applied to solve nonlinear systems Disadvantages: 1. May not converge or may converge slowly 2. Not appropriate for all systems Error bounds apply to solutions obtained by direct and iterative methods because they address the specification of [dA] and {db}. E. T. S. I. Caminos, Canales y Puertos
75
Iterative Solution Methods
Basic Mechanics: Starting with: a11x1 + a12x2 + a13x a1nxn = b1 a21x1 + a22x2 + a23x a2nxn = b2 a31x1 + a32x2 + a33x a3nxn = b3 : : an1x1 + an2x2 + an3x annxn = bn Solve each equation for one variable: x1 = [b1 – (a12x2 + a13x a1nxn )} / a11 x2 = [b2 – (a21x1 + a23x a2nxn )} / a22 x3 = [b3 – (a31x1 + a32x a3nxn )} / a33 : xn = [bn – (an1x2 + an2x an,n-1xn-1 )} / ann E. T. S. I. Caminos, Canales y Puertos
76
Iterative Solution Methods
Start with initial estimate of {x}0. Substitute into the right-hand side of all the equations. Generate new approximation {x}1. This is a multivariate one-point iteration: {x}j+1 = {g({x}j)} Repeat process until the maximum number of iterations reached or until: || xj+1 – xj || d + e || xj+1 || E. T. S. I. Caminos, Canales y Puertos
77
Convergence To solve [A]{x} = {b}
Separate [A] into: [A] = [Lo] + [D] + [Uo] [D] = diagonal (aii) [Lo] = lower triangular with 0's on diagonal [Uo] = upper triangular with 0's on diagonal Rewrite system: [A]{x} = ( [Lo] + [D] + [Uo] ){x} = {b} [D]{x} + ( [Lo] + [Uo] ){x} = {b} Iterate: [D]{x}j+1 = {b} – ( [Lo]+[Uo] ) {x}j {x}j+1 = [D]-1{b} – [D]-1 ( [Lo]+[Uo] ) {x}j Iterations converge if: || [D]-1 ( [Lo] + [Uo] ) || < 1 (sufficient if equations are diagonally dominant) E. T. S. I. Caminos, Canales y Puertos
78
Iterative Solution Methods – the Jacobi Method
E. T. S. I. Caminos, Canales y Puertos
79
Iterative Solution Methods -- Gauss-Seidel
In most cases using the newest values within the right-hand side equations will provide better estimates of the next value. If this is done, then we are using the Gauss-Seidel Method: ( [Lo]+[D] ){x}j+1 = {b} – [Uo] {x}j or explicitly: If this is done, then we are using the Gauss-Seidel Method E. T. S. I. Caminos, Canales y Puertos
80
Iterative Solution Methods -- Gauss-Seidel
If either method is going to converge, Gauss-Seidel will converge faster than Jacobi. Why use Jacobi at all? Because you can separate the n-equations into n independent tasks, it is very well suited computers with parallel processors. E. T. S. I. Caminos, Canales y Puertos
81
Convergence of Iterative Solution Methods
Rewrite given system: [A]{x} = { [B] + [E] } {x} = {b} where [B] is diagonal, or triangular so we can solve [B]{y} = {g} quickly. Thus, [B] {x}j+1= {b}– [E] {x}j which is effectively: {x}j+1 = [B]-1 ({b} – [E] {x}j ) True solution {x}c satisfies: {x}c = [B]-1 ({b} – [E] {x}c) Subtracting yields: {x}c – {x}j+1= – [B]-1 [E] [{x}c – {x}j] So ||{x}c – {x}j+1 || ||[B]-1 [E]|| ||{x}c – {x}j || Iterations converge linearly if || [B]-1 [E] || < 1 => || ([D] + [Lo])-1 [Uo] || < 1 For Gauss-Seidel => || [D] -1 ([Lo] + [Uo]) || < 1 For Jacobi E. T. S. I. Caminos, Canales y Puertos
82
Convergence of Iterative Solution Methods
Iterative methods will not converge for all systems of equations, nor for all possible rearrangements. If the system is diagonally dominant, i.e., | aii | > | aij | where i j then with all < 1.0, i.e., small slopes. E. T. S. I. Caminos, Canales y Puertos
83
Convergence of Iterative Solution Methods
A sufficient condition for convergence exists: Notes: 1. If the above does not hold, still may converge. 2. This looks similar to infinity norm of [A] E. T. S. I. Caminos, Canales y Puertos
84
Improving Rate of Convergence of G-S Iteration
Relaxation Schemes: where < l 2.0 (Usually the value of l is close to 1) Underrelaxation ( 0.0 < l < 1.0 ) More weight is placed on the previous value. Often used to: - make non-convergent system convergent or - to expedite convergence by damping out oscillations. Overrelaxation ( 1.0 < l 2.0 ) More weight is placed on the new value. Assumes that the new value is heading in right direction, and hence pushes new value close to true solution. The choice of l is highly problem-dependent and is empirical, so relaxation is usually only used for often repeated calculations of a particular class. E. T. S. I. Caminos, Canales y Puertos
85
Why Iterative Solutions?
We often need to solve [A]{x} = {b} where n = 1000's • Description of a building or airframe, • Finite-Difference approximations to PDE's. Most of A's elements will be zero; a finite-difference approximation to Laplace's equation will have five aij0 in each row of A. Direct method (Gaussian elimination) • Requires n3/3 flops (say n = 5000; n3/3 = 4 x 1010 flops) • Fills in many of n2-5n zero elements of A Iterative methods (Jacobi or Gauss-Seidel) • Never store [A] (say n = 5000; [A] would need 4n2 = 100 Mb) • Only need to compute [A-B] {x}; and to solve [B]{xt+1} = {b} E. T. S. I. Caminos, Canales y Puertos
86
Why Iterative Solutions?
• Effort: Suppose [B] is diagonal, solving [B] {v} = {b} n flops Computing [A-B] x 4n flops For m iterations 5mn flops For n = m = 5000, 5mn = 1.25x108 At worst O(n2). E. T. S. I. Caminos, Canales y Puertos
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.