SlideShare a Scribd company logo
1 of 31
(Managed By Shree Tapi Brahmcharyashram Sabha)
Shree Swami AtmanandVidhya Sankul, Kapodra,Varachha Road, Surat, Gujarat, India. 395006
Phone: 0261-2573552 Fax No.: 0261-2573554 Email: ssasit@yahoo.in Web: www.ssasit.ac.in
Shree Swami Atmanand Saraswati Institute of Technology,
Surat
1
Inner Product
Space
Presented by:
Patel Raj G.
Inner Product Spaces
Orthogonal and Orthogonal basis
Gram Schmidt Process
Orthogonal complements
 Orthogonal projections
Least Squares Approximation
Contents
Inner Product Spaces
(1)
(2)
(3)
(4)
(5) if and only if
〉〈〉〈 uvvu ,, 
〉〈〉〈〉〈 wuvuwvu ,,, 
〉,〈〉,〈 vuvu kk 
0, 〉〈 vv
0, 〉〈 vv v 0
 Inner product : represented by angle brackets
Let u, v, and w be vectors in a vector space V, and let c be
any scalar. An inner product on V is a function that associates
a real number with each pair of vectors u and v and
satisfies the following axioms (abstraction definition from
the properties of dot product in Theorem 5.3 on Slide 5.12)
(commutative property of the inner product)
(distributive property of the inner product
over vector addition)
(associative property of the scalar multiplication and the
inner product)
,u v〈 〉
,u v〈 〉
 Note:
dot product (Euclidean inner product for )
, general inner product for a vector space
n
R
V
 
 
u v
u v
 Note:
A vector space V with an inner product is called an inner
product space
Vector space:
Inner product space:
( , , )V  
( , , , , >)V   
 Properties of inner products
Let u, v, and w be vectors in an inner product space V, and
let c be any real number
(1)
(2)
(3)
, , 0 0 v v 0〈 〉〈 〉
, , ,  u v w u w v w〈 〉〈 〉〈 〉
, ,c cu v u v〈 〉 〈 〉
u and v are orthogonal if
 Distance between u and v:
 vuvuvuvu ,||||),(d
 Angle between two nonzero vectors u and v:
  0,
||||||||
,
cos
vu
vu 〉〈
 Orthogonal:
0, 〉〈 vu
( )u v
 Norm (length) of u:
〉〈 uuu ,|||| 
• The definition of norm (or length), distance, angle, orthogonal,
and normalizing for general inner product spaces closely parallel
to those based on the dot product in Euclidean n-space
 Properties of norm:
1)
(2) if and only if
(3)
 Properties of distance: (the same as the properties for the dot
product in R
n
on Slide 5.9)
(1)
(2) if and only if
(3)
0|||| u
0|||| u 0u 
|||||||||| uu cc 
0),( vud
0),( vud vu 
),(),( uvvu dd 
Let u and v be vectors in an inner product space V
(1) Cauchy-Schwarz inequality:
(2) Triangle inequality:
(3) Pythagorean theorem:
u and v are orthogonal if and only if
|||||||||||| vuvu 
222
|||||||||||| vuvu 
|||||||||,| vuvu 〉〈
 Orthogonal vectors:
Two vectors u and v in R
n
are orthogonal (perpendicular) if
0 vu
 Note:
The vector 0 is said to be orthogonal to every vector
Orthogonal and Orthogonal basis
 Ex : Finding orthogonal vectors
Determine all vectors in R
n
that are orthogonal to u = (4, 2)
0
24
),()2,4(
21
21



vv
vvvu
tv
t
v 

 21 ,
2

,
2
t
,t t R
 
   
 
v
)2,4(u Let ),( 21 vvv
Sol:
Orthogonal basis
 Definition:
a,b in V, ab if (a|b)=0.
 The zero vector is orthogonal to every vector.
An orthogonal set S is a set s.t. all pairs of distinct vectors
are orthogonal.
An orthonormal set S is an orthogonal set of unit vectors.
Every nonzero finite dimension inner product space has an
orthogonal basis
Theorems :
 If S={v1,v2,…,vn} is an orthogonal set of nonzero vectorin an
inner product space V then S is linearly independent.
 Any orthogonal set of n nonzero vector in Rn is basis for Rn.
 If S={v1,v2,…,vn} is an orthonormal basis for an inner product
space V, and u is any vector in V then it can be expressed as a
linear combination of v1,v2,…,vn.
u=<u,v1>v1 + <u,v2>v2 +…+ <u,vn>vn
 If S is an orthonormal basis for an n-dimensional inner product
space, and if coordinate vectors of u & v with respect to S are
[u]s=(a1,a2,…,an) and [v]s=(b1,b2,…,bn) then
||u||=√(a ) + (a ) + …+(a )
d(u,v)=√(a1-b1)2 + (a2-b2)2 +…+ (an-bn)2
<u,v>=a1b1 + a2b2 +…+ anbn
=[u]s.[v]s.
2 2 2
1 2
n
Gram Schmidt Process
 Gram schidt process to orthogonalisation of a set of vectors.
 Let {𝑢1, 𝑢2, 𝑢3} be the given set of vector which is basis for
vector space V.
 We shall constuct an orthogonal set{𝑣1 𝑣2 𝑣3} of vector of V
which becomes basis for V as under.
 Consider the vector space with the Euclidean inner product.
Apply the Gram–Schmidt process to transform the basis
vectors , , into an orthogonal basis ; then normalize the
orthogonal basis vectors to obtain an orthonormal basis .
 Step 1. Let 𝑉1=𝑈1
 Step 2. 𝑣2 = 𝑢2 −
< 𝑢2, 𝑣1> 𝑣1
||𝑣1||
2
 Step 3. 𝑣3 =𝑢3 −
< 𝑢3, 𝑣1> 𝑣1
|| 𝑣1||
2 −
< 𝑢3, 𝑣2> 𝑣2
|| 𝑣2||
2 and so on
……………………
 And orthonormal basis are
 {
𝑣1
||𝑣1||
,
𝑣2
||𝑣2||
,
𝑣3
||𝑣3||
}
Use gram schmidt process the set 𝑢1
=(1,1,1)𝑢2= (−1,1,0)𝑢3=(1,2,1)
 𝑣1 = 𝑢1= (1,1,1) ||𝑣1||2
= 1 + 1 + 1 = 3
 𝑢2 = −1,1,0
 <𝑢2, 𝑣1>=<(-1,1,0)(1,1,1)>
 =-1+1+0
=0
𝑣2 =𝑢2 −
< 𝑢2, 𝑣1> 𝑣1
||𝑣1||
2
=(-1,1,0)-
(0)(1,1,1)
3
𝑣2 = −1,1,0 ||𝑣2|| 2
=2
• Orthogonal basis set is
• {(1,1,1)(-1,1,0) (
1
6
,
1
6
, −
1
3
)}
• Now,orthonormal basis are
• {
𝑣1
||𝑣1||
,
𝑣2
||𝑣2||
,
𝑣3
||𝑣3||
}
• {(
1
√3
,
1
√3
,
1
√3
)(−
1
2
,
1
2
, 0)(
1
√6
,
1
√6
, −
2
√6
)}
Orthogonal complements
 Let W be a subspace of on inner product space V. A vector u
in V is orthogonal to W if is orthogonal to every vector in W.
The set of all vector in V that are orthogonal to W is called
orthogonal complement of W and is denoted by W.
Properties of Orthogonal complements
If W is a subspace of inner product space V than
● A vector u is in W if and only if u is orthogonal to every
spans W.
● The only vector common to W and W is 0.
● W is subspace of v.
● W W = W
● Let W=span {u ,u }
u =(2,0,-1)
u =(4,0,-2)
We have
The homo. system is
AX=0;











3
2
1
x
x
x
x
1
2
1 2
1
2








204
102
Ex : Fine the basis for an orthogonal complement of the subspace of
W span by the vector
u =(2,0,-1)
u =4,0,-2)










3
2
1
x
x
x









 
2
1
221
t
t
t
1
2
1
0
t
t 









2
1
0
21
t









 








0204
0102








0204
02/101
(1÷2)R1
The A.M. is
R2 +(-4)R1
~
~





 
0200
02/101
X = = = +
W = { ( 0,1,0), (1÷2,0,1)
 For inner product spaces:
Let u and v be two vectors in an inner product space V.
If , then the orthogonal projection of u onto v is
given by
0v 
,
proj
,
 

 
v
u v
u v
v v
2
Consider 0, cos
|| |||| || cos || |||| ||
|| |||| ||
proj
a a a a
a


   
 
  
  
    
 
v
v v v u
u v u v u v u v
v v u v v
u v u v u v
u v
v v v vv
u
v
proj , 0a a vu v

 Orthogonal projections : For the dot product function in R
n
, we
define the orthogonal projection of u onto v to be projvu = av (a
scalar multiple of v), and the coefficient a can be derived as
follows
 Ex : Finding an orthogonal projection in R3
Use the Euclidean inner product in R3
to find the
orthogonal projection of u = (6, 2, 4) onto v = (1, 2, 0)
Sol:
10)0)(4()2)(2()1)(6(,  vu
5021, 222
 vv
, 10
proj (1, 2 , 0) (2 , 4 , 0)
, 5
  
    
  
v
u v u v
u v v
v v v v
 Least Squares Approximation :
(A system of linear equations)bx A
11  mnnm
(1) When the system is consistent, we can use the Gaussian
elimination with the back substitution to solve for x
(2) When the system is inconsistent, only the “best possible”
solution of the system can be found, i.e., to find a solution of x
for which the difference (or said the error) between Ax and b is
smallest
Note: the system of linear equations Ax = b is consistent if and only if b is
in the column space of A
 Least Squares Approximation :
Given a system Ax = b of m linear equations in n unknowns,
the least squares problem is to find a vector x in Rn
that
minimizes the distance between Ax and b, i.e.,
with respect to the Euclidean inner product in R
n
. Such
vector is called a least squares solution of Ax = b
bxA
※ The term least squares comes from the fact that minimizing
is equivalent to minimizing = (Ax – b) ·
(Ax – b), which is a sum of squared errors
A x b
2
A x b
( )
ˆDefine ( ), and the problem to find
ˆsuch that is closest to is equivalent to find
the vector in ( ) closest to , that is proj
m n
n
W
A M
R
A CS A
W CS A
A
CS A




x
x
x
x b
b b
ˆThus proj
ˆ ˆ( proj ) ( ) ( ) ( )
ˆ ( ) ( )
ˆ( )
ˆ
W
W
Τ
Τ
Τ Τ
A
A W A CS A
A CS A NS A
A A
A A A


       
   
  
 
x b
b b b x b x
b x
b x 0
x b (the n×n linear system of normal equations associated with
Ax = b)
(The nullspace of AT is a solution space of
the homogeneous system ATx=0)
(To find the best solution which should satisfy this equation)ˆx
b
ˆAb x
ˆ projWAx = b
W
 Note:
The problem of finding the least squares solution of
is equal to the problem of finding an exact solution of the
associated normal system
bx A
bx ΤΤ
AAA ˆ
 Theorem associated with Least Squares Approximation solution:
For any linear system , the associated normal system
is consistent, and all solutions of the normal system are least
squares solution of Ax = b. In other words, if W is the column
space of A, and is the least squares solution of Ax = b, then
the orthogonal projection of b on W is , i.e.,
bx A
bx ΤΤ
AAA ˆ
xb ˆproj AW 
xˆ
ˆAx
 Ex : Solving the normal equations
Find the least squares solution of the following system
and find the orthogonal projection of b onto the column space
of A



























3
1
0
31
21
11
1
0
c
c
A bx
 Sol:
















































11
4
3
1
0
321
111
146
63
31
21
11
321
111
bT
T
A
AA
the corresponding normal system



















11
4
ˆ
ˆ
146
63
ˆ
1
0
c
c
AAA TT
bx
the least squares solution of Ax = b













2
3
3
5
1
0
ˆ
ˆ
ˆ
c
c
x
the orthogonal projection of b onto the column space of A





























6
17
6
8
6
1
2
3
3
5
)(
31
21
11
ˆproj xb AACS
Thanks

More Related Content

What's hot

Inner product space
Inner product spaceInner product space
Inner product spaceSheharBano31
 
ORTHOGONAL, ORTHONORMAL VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...
ORTHOGONAL, ORTHONORMAL  VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...ORTHOGONAL, ORTHONORMAL  VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...
ORTHOGONAL, ORTHONORMAL VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...Smit Shah
 
Independence, basis and dimension
Independence, basis and dimensionIndependence, basis and dimension
Independence, basis and dimensionATUL KUMAR YADAV
 
Linear dependence & independence vectors
Linear dependence & independence vectorsLinear dependence & independence vectors
Linear dependence & independence vectorsRakib Hossain
 
Chapter 4: Vector Spaces - Part 1/Slides By Pearson
Chapter 4: Vector Spaces - Part 1/Slides By PearsonChapter 4: Vector Spaces - Part 1/Slides By Pearson
Chapter 4: Vector Spaces - Part 1/Slides By PearsonChaimae Baroudi
 
Maths-->>Eigenvalues and eigenvectors
Maths-->>Eigenvalues and eigenvectorsMaths-->>Eigenvalues and eigenvectors
Maths-->>Eigenvalues and eigenvectorsJaydev Kishnani
 
Vector Spaces,subspaces,Span,Basis
Vector Spaces,subspaces,Span,BasisVector Spaces,subspaces,Span,Basis
Vector Spaces,subspaces,Span,BasisRavi Gelani
 
Linear transformation.ppt
Linear transformation.pptLinear transformation.ppt
Linear transformation.pptRaj Parekh
 
vector space and subspace
vector space and subspacevector space and subspace
vector space and subspace2461998
 
Eigen value and eigen vector
Eigen value and eigen vectorEigen value and eigen vector
Eigen value and eigen vectorRutvij Patel
 
Eigenvalues and eigenvectors
Eigenvalues and eigenvectorsEigenvalues and eigenvectors
Eigenvalues and eigenvectorsiraq
 
Diagonalization of Matrices
Diagonalization of MatricesDiagonalization of Matrices
Diagonalization of MatricesAmenahGondal1
 
Matrices and System of Linear Equations ppt
Matrices and System of Linear Equations pptMatrices and System of Linear Equations ppt
Matrices and System of Linear Equations pptDrazzer_Dhruv
 
linear equation and gaussian elimination
linear equation and gaussian eliminationlinear equation and gaussian elimination
linear equation and gaussian eliminationAju Thadikulangara
 

What's hot (20)

Inner product space
Inner product spaceInner product space
Inner product space
 
Vector Space.pptx
Vector Space.pptxVector Space.pptx
Vector Space.pptx
 
ORTHOGONAL, ORTHONORMAL VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...
ORTHOGONAL, ORTHONORMAL  VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...ORTHOGONAL, ORTHONORMAL  VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...
ORTHOGONAL, ORTHONORMAL VECTOR, GRAM SCHMIDT PROCESS, ORTHOGONALLY DIAGONALI...
 
Independence, basis and dimension
Independence, basis and dimensionIndependence, basis and dimension
Independence, basis and dimension
 
Linear dependence & independence vectors
Linear dependence & independence vectorsLinear dependence & independence vectors
Linear dependence & independence vectors
 
Diagonalization
DiagonalizationDiagonalization
Diagonalization
 
Chapter 4: Vector Spaces - Part 1/Slides By Pearson
Chapter 4: Vector Spaces - Part 1/Slides By PearsonChapter 4: Vector Spaces - Part 1/Slides By Pearson
Chapter 4: Vector Spaces - Part 1/Slides By Pearson
 
Maths-->>Eigenvalues and eigenvectors
Maths-->>Eigenvalues and eigenvectorsMaths-->>Eigenvalues and eigenvectors
Maths-->>Eigenvalues and eigenvectors
 
Maths
MathsMaths
Maths
 
Vector Spaces,subspaces,Span,Basis
Vector Spaces,subspaces,Span,BasisVector Spaces,subspaces,Span,Basis
Vector Spaces,subspaces,Span,Basis
 
Analytic function
Analytic functionAnalytic function
Analytic function
 
Linear transformation.ppt
Linear transformation.pptLinear transformation.ppt
Linear transformation.ppt
 
vector space and subspace
vector space and subspacevector space and subspace
vector space and subspace
 
Eigen value and eigen vector
Eigen value and eigen vectorEigen value and eigen vector
Eigen value and eigen vector
 
Eigenvalues and eigenvectors
Eigenvalues and eigenvectorsEigenvalues and eigenvectors
Eigenvalues and eigenvectors
 
Diagonalization of Matrices
Diagonalization of MatricesDiagonalization of Matrices
Diagonalization of Matrices
 
Ordinary differential equation
Ordinary differential equationOrdinary differential equation
Ordinary differential equation
 
Jacobians new
Jacobians newJacobians new
Jacobians new
 
Matrices and System of Linear Equations ppt
Matrices and System of Linear Equations pptMatrices and System of Linear Equations ppt
Matrices and System of Linear Equations ppt
 
linear equation and gaussian elimination
linear equation and gaussian eliminationlinear equation and gaussian elimination
linear equation and gaussian elimination
 

Similar to Inner Product Space

Innerproductspaces 151013072051-lva1-app6892 (1)
Innerproductspaces 151013072051-lva1-app6892 (1)Innerproductspaces 151013072051-lva1-app6892 (1)
Innerproductspaces 151013072051-lva1-app6892 (1)Himanshi Upadhyay
 
Vcla - Inner Products
Vcla - Inner ProductsVcla - Inner Products
Vcla - Inner ProductsPreetshah1212
 
Mathematical Foundations for Machine Learning and Data Mining
Mathematical Foundations for Machine Learning and Data MiningMathematical Foundations for Machine Learning and Data Mining
Mathematical Foundations for Machine Learning and Data MiningMadhavRao65
 
Vectors and Matrices: basis and dimension
Vectors and Matrices: basis and dimensionVectors and Matrices: basis and dimension
Vectors and Matrices: basis and dimensionMarry Chriselle Rañola
 
Vcla ppt ch=vector space
Vcla ppt ch=vector spaceVcla ppt ch=vector space
Vcla ppt ch=vector spaceMahendra Patel
 
Fixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsFixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsAlexander Decker
 
Chapter 4: Vector Spaces - Part 3/Slides By Pearson
Chapter 4: Vector Spaces - Part 3/Slides By PearsonChapter 4: Vector Spaces - Part 3/Slides By Pearson
Chapter 4: Vector Spaces - Part 3/Slides By PearsonChaimae Baroudi
 
Chapter 4: Vector Spaces - Part 5/Slides By Pearson
Chapter 4: Vector Spaces - Part 5/Slides By PearsonChapter 4: Vector Spaces - Part 5/Slides By Pearson
Chapter 4: Vector Spaces - Part 5/Slides By PearsonChaimae Baroudi
 
Presentation1 yash maths
Presentation1 yash mathsPresentation1 yash maths
Presentation1 yash mathsyash bhathawala
 
Presentation1 yash maths
Presentation1 yash mathsPresentation1 yash maths
Presentation1 yash mathsyash bhathawala
 
2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and DimensionCeni Babaoglu, PhD
 
Mat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptMat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptabidraufv
 
Orthogonal porjection in statistics
Orthogonal porjection in statisticsOrthogonal porjection in statistics
Orthogonal porjection in statisticsSahidul Islam
 
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear TransformationsCeni Babaoglu, PhD
 
Orthogonal Vector Spaces
Orthogonal Vector Spaces Orthogonal Vector Spaces
Orthogonal Vector Spaces Sohaib H. Khan
 

Similar to Inner Product Space (20)

Innerproductspaces 151013072051-lva1-app6892 (1)
Innerproductspaces 151013072051-lva1-app6892 (1)Innerproductspaces 151013072051-lva1-app6892 (1)
Innerproductspaces 151013072051-lva1-app6892 (1)
 
Vcla - Inner Products
Vcla - Inner ProductsVcla - Inner Products
Vcla - Inner Products
 
Mathematical Foundations for Machine Learning and Data Mining
Mathematical Foundations for Machine Learning and Data MiningMathematical Foundations for Machine Learning and Data Mining
Mathematical Foundations for Machine Learning and Data Mining
 
Vectors and Matrices: basis and dimension
Vectors and Matrices: basis and dimensionVectors and Matrices: basis and dimension
Vectors and Matrices: basis and dimension
 
150490106037
150490106037150490106037
150490106037
 
Vcla ppt ch=vector space
Vcla ppt ch=vector spaceVcla ppt ch=vector space
Vcla ppt ch=vector space
 
Inner product
Inner productInner product
Inner product
 
Fixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractionsFixed points theorem on a pair of random generalized non linear contractions
Fixed points theorem on a pair of random generalized non linear contractions
 
Chapter 4: Vector Spaces - Part 3/Slides By Pearson
Chapter 4: Vector Spaces - Part 3/Slides By PearsonChapter 4: Vector Spaces - Part 3/Slides By Pearson
Chapter 4: Vector Spaces - Part 3/Slides By Pearson
 
Chapter 4: Vector Spaces - Part 5/Slides By Pearson
Chapter 4: Vector Spaces - Part 5/Slides By PearsonChapter 4: Vector Spaces - Part 5/Slides By Pearson
Chapter 4: Vector Spaces - Part 5/Slides By Pearson
 
Ch4
Ch4Ch4
Ch4
 
Presentation1 yash maths
Presentation1 yash mathsPresentation1 yash maths
Presentation1 yash maths
 
Presentation1 yash maths
Presentation1 yash mathsPresentation1 yash maths
Presentation1 yash maths
 
2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension
 
Mat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptMat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.ppt
 
Vector space
Vector spaceVector space
Vector space
 
Chpt 2-sets v.3
Chpt 2-sets v.3Chpt 2-sets v.3
Chpt 2-sets v.3
 
Orthogonal porjection in statistics
Orthogonal porjection in statisticsOrthogonal porjection in statistics
Orthogonal porjection in statistics
 
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
 
Orthogonal Vector Spaces
Orthogonal Vector Spaces Orthogonal Vector Spaces
Orthogonal Vector Spaces
 

More from Patel Raj

disposal of solid waste
disposal of solid waste disposal of solid waste
disposal of solid waste Patel Raj
 
Communication Skill
Communication SkillCommunication Skill
Communication SkillPatel Raj
 
Water resources
Water  resourcesWater  resources
Water resourcesPatel Raj
 
Nanorobotics13.ppt
Nanorobotics13.pptNanorobotics13.ppt
Nanorobotics13.pptPatel Raj
 

More from Patel Raj (6)

disposal of solid waste
disposal of solid waste disposal of solid waste
disposal of solid waste
 
Hydrography
HydrographyHydrography
Hydrography
 
Array
ArrayArray
Array
 
Communication Skill
Communication SkillCommunication Skill
Communication Skill
 
Water resources
Water  resourcesWater  resources
Water resources
 
Nanorobotics13.ppt
Nanorobotics13.pptNanorobotics13.ppt
Nanorobotics13.ppt
 

Recently uploaded

CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learningmisbanausheenparvam
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
 
Software and Systems Engineering Standards: Verification and Validation of Sy...
Software and Systems Engineering Standards: Verification and Validation of Sy...Software and Systems Engineering Standards: Verification and Validation of Sy...
Software and Systems Engineering Standards: Verification and Validation of Sy...VICTOR MAESTRE RAMIREZ
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSKurinjimalarL3
 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLDeelipZope
 
microprocessor 8085 and its interfacing
microprocessor 8085  and its interfacingmicroprocessor 8085  and its interfacing
microprocessor 8085 and its interfacingjaychoudhary37
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx959SahilShah
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 

Recently uploaded (20)

CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
 
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learning
 
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Serviceyoung call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
 
Software and Systems Engineering Standards: Verification and Validation of Sy...
Software and Systems Engineering Standards: Verification and Validation of Sy...Software and Systems Engineering Standards: Verification and Validation of Sy...
Software and Systems Engineering Standards: Verification and Validation of Sy...
 
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCRCall Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCL
 
microprocessor 8085 and its interfacing
microprocessor 8085  and its interfacingmicroprocessor 8085  and its interfacing
microprocessor 8085 and its interfacing
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 

Inner Product Space

  • 1. (Managed By Shree Tapi Brahmcharyashram Sabha) Shree Swami AtmanandVidhya Sankul, Kapodra,Varachha Road, Surat, Gujarat, India. 395006 Phone: 0261-2573552 Fax No.: 0261-2573554 Email: ssasit@yahoo.in Web: www.ssasit.ac.in Shree Swami Atmanand Saraswati Institute of Technology, Surat 1 Inner Product Space
  • 3. Inner Product Spaces Orthogonal and Orthogonal basis Gram Schmidt Process Orthogonal complements  Orthogonal projections Least Squares Approximation Contents
  • 4. Inner Product Spaces (1) (2) (3) (4) (5) if and only if 〉〈〉〈 uvvu ,,  〉〈〉〈〉〈 wuvuwvu ,,,  〉,〈〉,〈 vuvu kk  0, 〉〈 vv 0, 〉〈 vv v 0  Inner product : represented by angle brackets Let u, v, and w be vectors in a vector space V, and let c be any scalar. An inner product on V is a function that associates a real number with each pair of vectors u and v and satisfies the following axioms (abstraction definition from the properties of dot product in Theorem 5.3 on Slide 5.12) (commutative property of the inner product) (distributive property of the inner product over vector addition) (associative property of the scalar multiplication and the inner product) ,u v〈 〉 ,u v〈 〉
  • 5.  Note: dot product (Euclidean inner product for ) , general inner product for a vector space n R V     u v u v  Note: A vector space V with an inner product is called an inner product space Vector space: Inner product space: ( , , )V   ( , , , , >)V   
  • 6.  Properties of inner products Let u, v, and w be vectors in an inner product space V, and let c be any real number (1) (2) (3) , , 0 0 v v 0〈 〉〈 〉 , , ,  u v w u w v w〈 〉〈 〉〈 〉 , ,c cu v u v〈 〉 〈 〉
  • 7. u and v are orthogonal if  Distance between u and v:  vuvuvuvu ,||||),(d  Angle between two nonzero vectors u and v:   0, |||||||| , cos vu vu 〉〈  Orthogonal: 0, 〉〈 vu ( )u v  Norm (length) of u: 〉〈 uuu ,||||  • The definition of norm (or length), distance, angle, orthogonal, and normalizing for general inner product spaces closely parallel to those based on the dot product in Euclidean n-space
  • 8.  Properties of norm: 1) (2) if and only if (3)  Properties of distance: (the same as the properties for the dot product in R n on Slide 5.9) (1) (2) if and only if (3) 0|||| u 0|||| u 0u  |||||||||| uu cc  0),( vud 0),( vud vu  ),(),( uvvu dd 
  • 9. Let u and v be vectors in an inner product space V (1) Cauchy-Schwarz inequality: (2) Triangle inequality: (3) Pythagorean theorem: u and v are orthogonal if and only if |||||||||||| vuvu  222 |||||||||||| vuvu  |||||||||,| vuvu 〉〈
  • 10.  Orthogonal vectors: Two vectors u and v in R n are orthogonal (perpendicular) if 0 vu  Note: The vector 0 is said to be orthogonal to every vector Orthogonal and Orthogonal basis
  • 11.  Ex : Finding orthogonal vectors Determine all vectors in R n that are orthogonal to u = (4, 2) 0 24 ),()2,4( 21 21    vv vvvu tv t v    21 , 2  , 2 t ,t t R         v )2,4(u Let ),( 21 vvv Sol:
  • 12. Orthogonal basis  Definition: a,b in V, ab if (a|b)=0.  The zero vector is orthogonal to every vector. An orthogonal set S is a set s.t. all pairs of distinct vectors are orthogonal. An orthonormal set S is an orthogonal set of unit vectors. Every nonzero finite dimension inner product space has an orthogonal basis
  • 13. Theorems :  If S={v1,v2,…,vn} is an orthogonal set of nonzero vectorin an inner product space V then S is linearly independent.  Any orthogonal set of n nonzero vector in Rn is basis for Rn.  If S={v1,v2,…,vn} is an orthonormal basis for an inner product space V, and u is any vector in V then it can be expressed as a linear combination of v1,v2,…,vn. u=<u,v1>v1 + <u,v2>v2 +…+ <u,vn>vn  If S is an orthonormal basis for an n-dimensional inner product space, and if coordinate vectors of u & v with respect to S are [u]s=(a1,a2,…,an) and [v]s=(b1,b2,…,bn) then ||u||=√(a ) + (a ) + …+(a ) d(u,v)=√(a1-b1)2 + (a2-b2)2 +…+ (an-bn)2 <u,v>=a1b1 + a2b2 +…+ anbn =[u]s.[v]s. 2 2 2 1 2 n
  • 14. Gram Schmidt Process  Gram schidt process to orthogonalisation of a set of vectors.  Let {𝑢1, 𝑢2, 𝑢3} be the given set of vector which is basis for vector space V.  We shall constuct an orthogonal set{𝑣1 𝑣2 𝑣3} of vector of V which becomes basis for V as under.  Consider the vector space with the Euclidean inner product. Apply the Gram–Schmidt process to transform the basis vectors , , into an orthogonal basis ; then normalize the orthogonal basis vectors to obtain an orthonormal basis .
  • 15.  Step 1. Let 𝑉1=𝑈1  Step 2. 𝑣2 = 𝑢2 − < 𝑢2, 𝑣1> 𝑣1 ||𝑣1|| 2  Step 3. 𝑣3 =𝑢3 − < 𝑢3, 𝑣1> 𝑣1 || 𝑣1|| 2 − < 𝑢3, 𝑣2> 𝑣2 || 𝑣2|| 2 and so on ……………………  And orthonormal basis are  { 𝑣1 ||𝑣1|| , 𝑣2 ||𝑣2|| , 𝑣3 ||𝑣3|| }
  • 16. Use gram schmidt process the set 𝑢1 =(1,1,1)𝑢2= (−1,1,0)𝑢3=(1,2,1)  𝑣1 = 𝑢1= (1,1,1) ||𝑣1||2 = 1 + 1 + 1 = 3  𝑢2 = −1,1,0  <𝑢2, 𝑣1>=<(-1,1,0)(1,1,1)>  =-1+1+0 =0 𝑣2 =𝑢2 − < 𝑢2, 𝑣1> 𝑣1 ||𝑣1|| 2 =(-1,1,0)- (0)(1,1,1) 3 𝑣2 = −1,1,0 ||𝑣2|| 2 =2
  • 17. • Orthogonal basis set is • {(1,1,1)(-1,1,0) ( 1 6 , 1 6 , − 1 3 )} • Now,orthonormal basis are • { 𝑣1 ||𝑣1|| , 𝑣2 ||𝑣2|| , 𝑣3 ||𝑣3|| } • {( 1 √3 , 1 √3 , 1 √3 )(− 1 2 , 1 2 , 0)( 1 √6 , 1 √6 , − 2 √6 )}
  • 18. Orthogonal complements  Let W be a subspace of on inner product space V. A vector u in V is orthogonal to W if is orthogonal to every vector in W. The set of all vector in V that are orthogonal to W is called orthogonal complement of W and is denoted by W.
  • 19. Properties of Orthogonal complements If W is a subspace of inner product space V than ● A vector u is in W if and only if u is orthogonal to every spans W. ● The only vector common to W and W is 0. ● W is subspace of v. ● W W = W
  • 20. ● Let W=span {u ,u } u =(2,0,-1) u =(4,0,-2) We have The homo. system is AX=0;            3 2 1 x x x x 1 2 1 2 1 2         204 102 Ex : Fine the basis for an orthogonal complement of the subspace of W span by the vector u =(2,0,-1) u =4,0,-2)
  • 21.           3 2 1 x x x            2 1 221 t t t 1 2 1 0 t t           2 1 0 21 t                    0204 0102         0204 02/101 (1÷2)R1 The A.M. is R2 +(-4)R1 ~ ~        0200 02/101 X = = = + W = { ( 0,1,0), (1÷2,0,1)
  • 22.  For inner product spaces: Let u and v be two vectors in an inner product space V. If , then the orthogonal projection of u onto v is given by 0v  , proj ,      v u v u v v v 2 Consider 0, cos || |||| || cos || |||| || || |||| || proj a a a a a                      v v v v u u v u v u v u v v v u v v u v u v u v u v v v v vv u v proj , 0a a vu v   Orthogonal projections : For the dot product function in R n , we define the orthogonal projection of u onto v to be projvu = av (a scalar multiple of v), and the coefficient a can be derived as follows
  • 23.  Ex : Finding an orthogonal projection in R3 Use the Euclidean inner product in R3 to find the orthogonal projection of u = (6, 2, 4) onto v = (1, 2, 0) Sol: 10)0)(4()2)(2()1)(6(,  vu 5021, 222  vv , 10 proj (1, 2 , 0) (2 , 4 , 0) , 5            v u v u v u v v v v v v
  • 24.  Least Squares Approximation : (A system of linear equations)bx A 11  mnnm (1) When the system is consistent, we can use the Gaussian elimination with the back substitution to solve for x (2) When the system is inconsistent, only the “best possible” solution of the system can be found, i.e., to find a solution of x for which the difference (or said the error) between Ax and b is smallest Note: the system of linear equations Ax = b is consistent if and only if b is in the column space of A
  • 25.  Least Squares Approximation : Given a system Ax = b of m linear equations in n unknowns, the least squares problem is to find a vector x in Rn that minimizes the distance between Ax and b, i.e., with respect to the Euclidean inner product in R n . Such vector is called a least squares solution of Ax = b bxA ※ The term least squares comes from the fact that minimizing is equivalent to minimizing = (Ax – b) · (Ax – b), which is a sum of squared errors A x b 2 A x b
  • 26. ( ) ˆDefine ( ), and the problem to find ˆsuch that is closest to is equivalent to find the vector in ( ) closest to , that is proj m n n W A M R A CS A W CS A A CS A     x x x x b b b ˆThus proj ˆ ˆ( proj ) ( ) ( ) ( ) ˆ ( ) ( ) ˆ( ) ˆ W W Τ Τ Τ Τ A A W A CS A A CS A NS A A A A A A                    x b b b b x b x b x b x 0 x b (the n×n linear system of normal equations associated with Ax = b) (The nullspace of AT is a solution space of the homogeneous system ATx=0) (To find the best solution which should satisfy this equation)ˆx b ˆAb x ˆ projWAx = b W
  • 27.  Note: The problem of finding the least squares solution of is equal to the problem of finding an exact solution of the associated normal system bx A bx ΤΤ AAA ˆ  Theorem associated with Least Squares Approximation solution: For any linear system , the associated normal system is consistent, and all solutions of the normal system are least squares solution of Ax = b. In other words, if W is the column space of A, and is the least squares solution of Ax = b, then the orthogonal projection of b on W is , i.e., bx A bx ΤΤ AAA ˆ xb ˆproj AW  xˆ ˆAx
  • 28.  Ex : Solving the normal equations Find the least squares solution of the following system and find the orthogonal projection of b onto the column space of A                            3 1 0 31 21 11 1 0 c c A bx
  • 30. the least squares solution of Ax = b              2 3 3 5 1 0 ˆ ˆ ˆ c c x the orthogonal projection of b onto the column space of A                              6 17 6 8 6 1 2 3 3 5 )( 31 21 11 ˆproj xb AACS