Download presentation
Presentation is loading. Please wait.
1
Orthogonal Projection
Hung-yi Lee
2
Reference Textbook: Chapter 7.3, 7.4
3
Orthogonal Projection
What is Orthogonal Complement What is Orthogonal Projection How to do Orthogonal Projection Application of Orthogonal Projection
4
Orthogonal Complement
The orthogonal complement of a nonempty vector set S is denoted as S (S perp). S is the set of vectors that are orthogonal to every vector in S 𝑆 = 𝑣:𝑣⋅𝑢=0,∀𝑢∈𝑆 W S = Rn S = {0} W S = {0} S = Rn
5
Orthogonal Complement
The orthogonal complement of a nonempty vector set S is denoted as S (S perp). S is the set of vectors that are orthogonal to every vector in S 𝑆 = 𝑣:𝑣⋅𝑢=0,∀𝑢∈𝑆 V W: 𝑊= 𝑤 1 𝑤 | 𝑤 1 , 𝑤 2 ∈R by z e1 = z e2 = 0 for all v V and w W, v w = 0 W V: 𝑉= 𝑣 3 | 𝑣 3 ∈R = W? since e1, e2 W, all z = [ z1 z2 z3 ]T W must have z1 = z2 = 0
6
Properties of Orthogonal Complement
Is 𝑆 always a subspace? For any nonempty vector set S, 𝑆𝑝𝑎𝑛 𝑆 = 𝑆 Let W be a subspace, and B be a basis of W. Use the three propertites 𝐵 = 𝑊 What is 𝑆∩𝑆 ? Zero vector
7
Properties of Orthogonal Complement
Example: For W = Span{u1, u2}, where u1 = [ 1 4 ]T and u2 =[ 1 ]T v W if and only if u1 v = u2 v = 0 i.e., v = [ x1 x2 x3 x4 ]T satisfies Proof You show that (a basis of W)(a basis of W) = a basis of Rn. is a basis for W. W = Solutions of “Ax=0” = Null A
8
Properties of Orthogonal Complement
For any matrix A 𝑅𝑜𝑤 𝐴 =𝑁𝑢𝑙𝑙 𝐴 v (Row A) For all w Span{rows of A}, w v = 0 Av = 0. C𝑜𝑙 𝐴 =𝑁𝑢𝑙𝑙 𝐴 𝑇 (Col A) = (Row AT) = Null AT . For any subspace W of Rn 𝑑𝑖𝑚𝑊+𝑑𝑖𝑚 𝑊 =𝑛 rank nullity
9
Unique For any subspace W of Rn 𝑑𝑖𝑚𝑊+𝑑𝑖𝑚 𝑊 =𝑛
Basis: 𝑤 1 , 𝑤 2 ,⋯, 𝑤 𝑘 Basis: 𝑧 1 , 𝑧 2 ,⋯, 𝑧 𝑛−𝑘 Basis for Rn For every vector u, W u u = w + z (unique) z W w ∈𝑊 ∈ 𝑊
10
Orthogonal Projection
What is Orthogonal Complement What is Orthogonal Projection How to do Orthogonal Projection Application of Orthogonal Projection
11
Orthogonal Projection
u = w + z (unique) ∈𝑊 ∈ 𝑊 ∈ 𝑊 𝑧 Orthogonal Projection Operator: The function 𝑈 𝑊 𝑢 is the orthogonal projection of u on W. Linear?
12
Orthogonal Projection
w in subspace W is “closest” to vector u. ∈ 𝑊 𝑧 “closest” Always orthogonal z is always orthogonal to all vectors in W.
13
Closest Vector Property
Among all vectors in subspace W, the vector closest to u is the orthogonal projection of u on W u w, w, w w W. (u w) (w w) = 0. u w2 = (u w) + (w w)2 = u w2 + w w2 > u w2 W w w’ w = UW(u) The distance from a vector u to a subspace W is the distance between u and the orthogonal projection of u on W
14
Orthogonal Projection Matrix
Orthogonal projection operator is linear. It has standard matrix. Orthogonal Projection Matrix Pw u z = u - w Proposition: UW is linear. Proof If UW(u1) = w1 and UW(u2) = w2, then unique z1, z2 W such that u1 = w1+z1 and u2 = w2+z2 u1+u2 = w1+w2 + z1+z2 UW(u1+u2) = w1+w2 since w1+w2 W and z1+z2 W Similarly UW(cu) = cw c, u = 𝑃 𝑊 𝑢 w = UW(u) v = UW(v) = 𝑃 𝑊 𝑣 W
15
Orthogonal Projection
What is Orthogonal Complement What is Orthogonal Projection How to do Orthogonal Projection Application of Orthogonal Projection
16
Orthogonal Projection on a line
Orthogonal projection of a vector on a line 𝑧∙𝑢=0 v v: any vector u: any nonzero vector on L w: orthogonal projection of v onto L , w = cu z: v w z L z u w W in terms of u Tip: 頂端;尖端 𝑣−𝑤 ∙𝑢 = 𝑣−𝑐𝑢 ∙𝑢 =𝑣∙𝑢−𝑐𝑢∙𝑢 =𝑣∙𝑢−𝑐 𝑢 2 =0 𝑐= 𝑣∙𝑢 𝑢 2 𝑤=𝑐𝑢= 𝑣∙𝑢 𝑢 2 𝑢 = 𝑣− 𝑣∙𝑢 𝑢 2 𝑢 Distance from tip of v to L : 𝑧 = 𝑣−𝑤
17
Orthogonal Projection
𝑐= 𝑣∙𝑢 𝑢 2 𝑤=𝑐𝑢= 𝑣∙𝑢 𝑢 2 𝑢 Example: L u v w z 𝑧∙𝑢=0 L is y = (1/2)x 𝑣= 4 1 𝑢= 2 1 𝑤= 𝑣∙𝑢 𝑢 2 𝑢 = 4 1 − = −4 =
18
Orthogonal Projection Matrix
Let C be an n x k matrix whose columns form a basis for a subspace W 𝑃 𝑊 =𝐶 𝐶 𝑇 𝐶 −1 𝐶 𝑇 n x n Proof: Let u Rn and w = UW(u). Since W = Col C, w = Cv for some v Rk and u w W 0 = CT(u w) = CTu CTw = CTu CTCv. CTu = CTCv. v = (CTC)1CTu and w = C(CTC)1CTu as CTC is invertible.
19
Orthogonal Projection Matrix
Let C be an n x k matrix whose columns form a basis for a subspace W n x n 𝑃 𝑊 =𝐶 𝐶 𝑇 𝐶 −1 𝐶 𝑇 Let C be a matrix with linearly independent columns. Then 𝐶 𝑇 𝐶 is invertible. Proof: We want to prove that CTC has independent columns. Suppose CTCb = 0 for some b. bTCTCb = (Cb)TCb = (Cb) (Cb) = Cb2 = 0. Cb = 0 b = 0 since C has L.I. columns. Thus CTC is invertible.
20
Orthogonal Projection Matrix
Example: Let W be the 2-dimensional subspace of R3 with equation x1 x2 +2x3 = 0. 𝑃 𝑊 =𝐶 𝐶 𝑇 𝐶 −1 𝐶 𝑇 W has a basis 𝐶= 𝑃 𝑊 = 𝑃 𝑊 =
21
Orthogonal Projection
What is Orthogonal Complement What is Orthogonal Projection How to do Orthogonal Projection Application of Orthogonal Projection
22
Solution of Inconsistent System of Linear Equations
Suppose Ax = b is an inconsistent system of linear equations. b is not in the column space of A Find vector z minimizing ||Az b|| b ||Ax b|| ||Az b|| Ax Az = PWb W = Col A
23
Least Square Approximation
data pairs: x1 y1 x2 y2 xi yi predict e.g. Regression回歸 (今天股票,明天股票) (今天PM2.5,明天PM2.5) Find the “least-square line” y = a0 + a1x to best fit the data Regression
24
Least Square Approximation
𝑦= 𝑎 0 + 𝑎 1 𝑥 Error Vector: 𝑥 𝑖 , 𝑎 0 + 𝑎 1 𝑥 𝑖 𝑦 𝑖 − 𝑎 0 + 𝑎 1 𝑥 𝑖 𝑥 𝑖 , 𝑦 𝑖 Find a0 and a1 minimizing E
25
Least Square Approximation
Error Vector: Find a0 and a1 minimizing E E = y (a0v1 + a1v2)2 = y Ca2
26
Least Square Approximation
Find a minimizing E = y Ca2 B = {v1, v2} (L.I.) Ca is the orthogonal projection of y on W = Span B . find a such that Ca = PWy
27
Example 1 y = 0.056 + 0.745x. Prediction:
connecting rod Prediction: if the rough weight is 2.65, the finished weight is (2.65) = (estimation)
28
Least Square Approximation
Best quadratic fit: using y = a0 + a1x + a2x2 to fit the data points (x1, y1), (x2, y2), , (xn, yn) 𝑒= 𝑦 1 − 𝑎 0 + 𝑎 1 𝑥 1 + 𝑎 2 𝑥 𝑦 2 − 𝑎 0 + 𝑎 1 𝑥 2 + 𝑎 2 𝑥 ⋮ 𝑦 𝑛 − 𝑎 0 + 𝑎 1 𝑥 𝑛 + 𝑎 2 𝑥 𝑛 2 y = a0 + a1x + a2x2 Find a0, a1 and a2 minimizing E
29
Least Square Approximation
Best quadratic fit: using y = a0 + a1x + a2x2 to fit the data points (x1, y1), (x2, y2), , (xn, yn) 𝑒= 𝑦 1 − 𝑎 0 + 𝑎 1 𝑥 1 + 𝑎 2 𝑥 𝑦 2 − 𝑎 0 + 𝑎 1 𝑥 2 + 𝑎 2 𝑥 ⋮ 𝑦 𝑛 − 𝑎 0 + 𝑎 1 𝑥 𝑛 + 𝑎 2 𝑥 𝑛 2 Find a0, a1 and a2 minimizing E
30
y = a0 + a1x + a2x2 𝑦= 𝑥−16.11 𝑥 2 Best fitting polynomial of any desired maximum degree may be found with the same method.
31
Multivariable Least Square Approximation
𝑦 𝐴 𝑦 𝐴 = 𝑎 0 + 𝑎 1 𝑥 𝐴 + 𝑎 2 𝑥 𝐵 𝑥 𝐴 𝑥 𝐵
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.