Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen.

Slides:



Advertisements
Similar presentations
10.4 Complex Vector Spaces.
Advertisements

Chapter 4 Euclidean Vector Spaces
Chapter 5 Inner Product Spaces
6 6.1 © 2012 Pearson Education, Inc. Orthogonality and Least Squares INNER PRODUCT, LENGTH, AND ORTHOGONALITY.
Chapter 5 Orthogonality
CHAPTER 5 INNER PRODUCT SPACES
Orthogonality and Least Squares
6 6.3 © 2012 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
Chapter 5 Inner Product Spaces
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
1 MAC 2103 Module 10 lnner Product Spaces I. 2 Rev.F09 Learning Objectives Upon completing this module, you should be able to: 1. Define and find the.
Length and Dot Product in R n Notes: is called a unit vector. Notes: The length of a vector is also called its norm. Chapter 5 Inner Product Spaces.
Chapter 9-Vectors Calculus, 2ed, by Blank & Krantz, Copyright 2011 by John Wiley & Sons, Inc, All Rights Reserved.
Elementary Linear Algebra Howard Anton Copyright © 2010 by John Wiley & Sons, Inc. All rights reserved. Chapter 3.
1 February 24 Matrices 3.2 Matrices; Row reduction Standard form of a set of linear equations: Chapter 3 Linear Algebra Matrix of coefficients: Augmented.
8.1 Vector spaces A set of vector is said to form a linear vector space V Chapter 8 Matrices and vector spaces.
Inner Product Spaces Euclidean n-space: Euclidean n-space: vector lengthdot productEuclidean n-space R n was defined to be the set of all ordered.
Chapter 5: The Orthogonality and Least Squares
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
Chapter 5 Orthogonality.
Vectors in R n a sequence of n real number An ordered n-tuple: the set of all ordered n-tuple  n-space: R n Notes: (1) An n-tuple can be viewed.
Vectors CHAPTER 7. Ch7_2 Contents  7.1 Vectors in 2-Space 7.1 Vectors in 2-Space  7.2 Vectors in 3-Space 7.2 Vectors in 3-Space  7.3 Dot Product 7.3.
AN ORTHOGONAL PROJECTION
Chap. 6 Linear Transformations
1 Chapter 5 – Orthogonality and Least Squares Outline 5.1 Orthonormal Bases and Orthogonal Projections 5.2 Gram-Schmidt Process and QR Factorization 5.3.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Chapter 4 Linear Transformations 4.1 Introduction to Linear Transformations 4.2 The Kernel and Range of a Linear Transformation 4.3 Matrices for Linear.
Lecture 11 Inner Product Space
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
Section 5.1 Length and Dot Product in ℝ n. Let v = ‹v 1­­, v 2, v 3,..., v n › and w = ‹w 1­­, w 2, w 3,..., w n › be vectors in ℝ n. The dot product.
Chap. 5 Inner Product Spaces 5.1 Length and Dot Product in R n 5.2 Inner Product Spaces 5.3 Orthonormal Bases: Gram-Schmidt Process 5.4 Mathematical Models.
Lecture 13 Inner Product Space & Linear Transformation Last Time - Orthonormal Bases:Gram-Schmidt Process - Mathematical Models and Least Square Analysis.
Chap. 4 Vector Spaces 4.1 Vectors in Rn 4.2 Vector Spaces
7.1 Eigenvalues and Eigenvectors
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
1 MAC 2103 Module 11 lnner Product Spaces II. 2 Rev.F09 Learning Objectives Upon completing this module, you should be able to: 1. Construct an orthonormal.
Lecture 9 Vector & Inner Product Space
6 6.3 © 2016 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
Lecture 9 Vector & Inner Product Space Last Time - Spanning Sets and Linear Independence (Cont.) - Basis and Dimension - Rank of a Matrix and Systems of.
Lecture 7 Vector Spaces Lat Time Properties of Determinants Introduction to Eigenvalues Applications of Determinants Vectors in R n Elementary Linear Algebra.
Matrices, Vectors, Determinants.
Chapter 4 Linear Transformations 4.1 Introduction to Linear Transformations 4.2 The Kernel and Range of a Linear Transformation 4.3 Matrices for Linear.
Lecture 11 Inner Product Space Last Time - Coordinates and Change of Basis - Applications - Length and Dot Product in R n Elementary Linear Algebra R.
Inner Product Spaces Euclidean n-space: Euclidean n-space: vector lengthdot productEuclidean n-space R n was defined to be the set of all ordered.
Lecture 9 Vector & Inner Product Spaces Last Time Spanning Sets and Linear Independence (Cont.) Basis and Dimension Rank of a Matrix and Systems of Linear.
EE611 Deterministic Systems Vector Spaces and Basis Changes Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Lecture 7 Vector Space Last Time - Properties of Determinants
Review of Linear Algebra
Lecture 2 Matrices Lat Time - Course Overview
Chapter 1 Linear Equations and Vectors
Elementary Linear Algebra
Matrices and vector spaces
Lecture 10 Inner Product Spaces
Lecture 8 Vector Space Last Time - Vector Spaces and Applications
Lecture 7 Vector Space Last Time - Properties of Determinants
Lecture 14 Linear Transformation
Orthogonality and Least Squares
Chapter 3 Linear Algebra
Linear Algebra Lecture 39.
Linear Algebra Lecture 38.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Linear Algebra Lecture 20.
Chapter 4 Linear Transformations
RAYAT SHIKSHAN SANSTHA’S S.M.JOSHI COLLEGE HADAPSAR, PUNE
Null Spaces, Column Spaces, and Linear Transformations
NULL SPACES, COLUMN SPACES, AND LINEAR TRANSFORMATIONS
Orthogonality and Least Squares
Chapter 5 Inner Product Spaces
Presentation transcript:

Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen et al. (6th Edition) TKUEE 翁慶昌 -NTUEE SCC_12_2008

Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project Reading Assignment: Secs

5.2 Inner Product Spaces (1) (2) (3) (4) and if and only if Inner product: Let u, v, and w be vectors in a vector space V, and let c be any scalar. An inner product on V is a function that associates a real number with each pair of vectors u and v and satisfies the following axioms

Note: A vector space V with an inner product is called an inner product space. Vector space: Inner product space:

Thm 5.7: (Properties of inner products) Let u, v, and w be vectors in an inner product space V, and let c be any real number. (1) (2) (3) Norm (length) of u: Note:

u and v are orthogonal if. Distance between u and v: Angle between two nonzero vectors u and v: Orthogonal:

Notes: (1) If, then v is called a unit vector. (2) (the unit vector in the direction of v) not a unit vector

Ex 6: (Finding inner product) is an inner product Sol:

Properties of norm: (1) (2) if and only if (3) Properties of distance: (1) (2) if and only if (3)

Thm 5.8 : Let u and v be vectors in an inner product space V. (1) Cauchy-Schwarz inequality: (2) Triangle inequality: (3) Pythagorean theorem : u and v are orthogonal if and only if Theorem 5.5 Theorem 5.6 Theorem 5.4

Orthogonal projections in inner product spaces: Let u and v be two vectors in an inner product space V, such that. Then the orthogonal projection of u onto v is given by Note: If v is a init vector, then. The formula for the orthogonal projection of u onto v takes the following simpler form. Q: Why called orthogonal projection?

Ex 10: (Finding an orthogonal projection in R 3 ) Use the Euclidean inner product in R 3 to find the orthogonal projection of u=(6, 2, 4) onto v=(1, 2, 0). Sol:

Thm 5.9: (Orthogonal projection and distance) Let u and v be two vectors in an inner product space V, such that. Then

Keywords in Section 5.2: inner product: 內積 inner product space: 內積空間 norm: 範數 distance: 距離 angle: 夾角 orthogonal: 正交 unit vector: 單位向量 normalizing: 單範化 Cauchy – Schwarz inequality: 科西 - 舒瓦茲不等式 triangle inequality: 三角不等式 Pythagorean theorem: 畢氏定理 orthogonal projection: 正交投影

Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

Orthonormal Bases: Gram-Schmidt Process Orthogonal: A set S of vectors in an inner product space V is called an orthogonal set if every pair of vectors in the set is orthogonal. Orthonormal: An orthogonal set in which each vector is a unit vector is called orthonormal. Note: If S is a basis, then it is called an orthogonal basis or an orthonormal basis.

Ex 1: (A nonstandard orthonormal basis for R 3 ) Show that the following set is an orthonormal basis. Sol: Show that the three vectors are mutually orthogonal.

Show that each vector is of length 1. Thus S is an orthonormal set.

The standard basis is orthonormal. Ex 2: (An orthonormal basis for ) In, with the inner product Sol: Then

Thm 5.10: (Orthogonal sets are linearly independent) If is an orthogonal set of nonzero vectors in an inner product space V, then S is linearly independent. Pf: S is an orthogonal set of nonzero vectors

Corollary to Thm 5.10: If V is an inner product space of dimension n, then any orthogonal set of n nonzero vectors is a basis for V.

Ex 4: (Using orthogonality to test for a basis) Show that the following set is a basis for. Sol: : nonzero vectors (by Corollary to Theorem 5.10)

Thm 5.11: (Coordinates relative to an orthonormal basis) If is an orthonormal basis for an inner product space V, then the coordinate representation of a vector w with respect to B is is orthonormal ( unique representation ) Pf: is a basis for V

Note: If is an orthonormal basis for V and, Then the corresponding coordinate matrix of w relative to B is

Ex 5: (Representing vectors relative to an orthonormal basis) Find the coordinates of w = (5, -5, 2) relative to the following orthonormal basis for. Sol:

Gram-Schmidt orthonormalization process: is a basis for an inner product space V is an orthogonal basis. is an orthonormal basis.

Sol: Ex 7: (Applying the Gram-Schmidt orthonormalization process) Apply the Gram-Schmidt process to the following basis.

Orthogonal basis Orthonormal basis

Ex 10: (Alternative form of Gram-Schmidt orthonormalization process) Find an orthonormal basis for the solution space of the homogeneous system of linear equations. Sol:

Thus one basis for the solution space is ( orthogonal basis) (orthonormal basis )

Keywords in Section 5.3: orthogonal set: 正交集合 orthonormal set: 單範正交集合 orthogonal basis: 正交基底 orthonormal basis: 單範正交基底 linear independent: 線性獨立 Gram-Schmidt Process: Gram-Schmidt 過程

Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

Diffusion of Innovations: Bass Model S(t)=m[p+qF(t)][(1-F(t)] m=ultimate market potential p=coefficient of innovation q=coefficient of imitation

(c) Frank M. Bass (1999) An Empirical Generalization

(c) Frank M. Bass (1999) Another Example 35 mm Projectors

(c) Frank M. Bass (1999) Another Example: Overhead Projectors

Mathematical Models and Least Squares Analysis Let W be a subspace of an inner product space V. (a) A vector u in V is said to orthogonal to W, if u is orthogonal to every vector in W. (b) The set of all vectors in V that are orthogonal to W is called the orthogonal complement of W. (read “ perp”) Orthogonal complement of W:  Notes:

Notes:  Ex:

Direct sum: Let and be two subspaces of. If each vector can be uniquely written as a sum of a vector from and a vector from,, then is the direct sum of and, and you can write. Thm 5.13: (Properties of orthogonal subspaces) Let W be a subspace of R n. Then the following properties are true. (1) (2) (3)

Thm 5.14: (Projection onto a subspace) If is an orthonormal basis for the subspace W of V, and, then

Ex 5: (Projection onto a subspace) Find the projection of the vector v onto the subspace W. Sol: an orthogonal basis for W an orthonormal basis for W

Find by the other method:

Thm 5.15: (Orthogonal projection and distance) Let W be a subspace of an inner product space V, and. Then for all, ( is the best approximation to v from W)

Pf: By the Pythagorean theorem

Notes: (1) Among all the scalar multiples of a vector u, the orthogonal projection of v onto u is the one that is closest to v. (2) Among all the vectors in the subspace W, the vector is the closest vector to v.

Thm 5.16: (Fundamental subspaces of a matrix) If A is an m×n matrix, then (1) (2) (3) (4)

Ex 6: (Fundamental subspaces) Find the four fundamental subspaces of the matrix. (reduced row-echelon form) Sol:

Check:

Ex 3: Let W is a subspace of R 4 and. (a) Find a basis for W (b) Find a basis for the orthogonal complement of W. Sol: (reduced row-echelon form)

is a basis for W Notes:

Least squares problem: (A system of linear equations) (1) When the system is consistent, we can use the Gaussian elimination with back-substitution to solve for x (2) When the system is consistent, how to find the “best possible” solution of the system. That is, the value of x for which the difference between Ax and b is small.

Least squares solution: Given a system Ax = b of m linear equations in n unknowns, the least squares problem is to find a vector x in R n that minimizes with respect to the Euclidean inner product on R n. Such a vector is called a least squares solution of Ax = b.

(the normal system associated with Ax = b)

Note: The problem of finding the least squares solution of is equal to he problem of finding an exact solution of the associated normal system. Thm: For any linear system, the associated normal system is consistent, and all solutions of the normal system are least squares solution of Ax = b. Moreover, if W is the column space of A, and x is any least squares solution of Ax = b, then the orthogonal projection of b on W is

Thm: If A is an m×n matrix with linearly independent column vectors, then for every m×1 matrix b, the linear system Ax = b has a unique least squares solution. This solution is given by Moreover, if W is the column space of A, then the orthogonal projection of b on W is

Ex 7: (Solving the normal equations) Find the least squares solution of the following system and find the orthogonal projection of b on the column space of A.

Sol: the associated normal system

the least squares solution of Ax = b the orthogonal projection of b on the column space of A

Keywords in Section 5.4: orthogonal to W: 正交於 W orthogonal complement: 正交補集 direct sum: 直和 projection onto a subspace: 在子空間的投影 fundamental subspaces: 基本子空間 least squares problem: 最小平方問題 normal equations: 一般方程式

Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project

/7/9 Application: Cross Product Cross product (vector product) of two vectors  向量 (vector) 方向 : use right-hand rule The cross product is not commutative: The cross product is distributive:

/7/9 Parallelogram representation of the vector product x y θ Bsinθ Area Application: Cross Product

/7/9 向量之三重純量積 Triple Scalar product The dot and the cross may be interchanged : 純量 (scalar)

/7/9 向量之三重純量積 Parallelepiped representation of triple scalar product x y z Volume of parallelepiped defined by,, and

Fourier Approximation

Fourier Approximation The Fourier series transforms a given periodic function into a superposition of sine and cosine waves The following equations are used

Lecture 11: Inner Product Spaces Today Inner Product Spaces (Cont.) Orthonormal Bases:Gram-Schmidt Process Mathematical Models and Least Squares Analysis Applications & Term Project Reading Assignment: Secs

Lecture 11: Inner Product Space Next Time Introduction to Linear Transformations The Kernel and Range of a Linear Transformation Matrices for Linear Transformations Reading Assignment: Secs