SlideShare a Scribd company logo
1 of 6
Download to read offline
Hi please complete the following with detailed working out Find the eigenvalues of C1 C1 C1
C1 C2 C3 C3 C3 C3 cn cn Hint: You may find that the identity A2 (c1 c2 cm) is useful.
Solution
COORDINATES, EIGENVALUES, AND EIGENVECTORS
1. Review of coordinates
1.1. The standard basis. The standard basis for R n is the basis E = {e1, e2, . . . en}, where ej is
the vector with 1 in the j-th position and zeros elsewhere.
1.2. Coordinates with respect to a basis. Let B = {v1, v2, . . . vn} be a basis for R n .
We know that for each vector x R n there are unique scalar “coordinates” c1, c2, . . . cn such that
(1.1) x = c1v1 + c2v2 + . . . + cnvn
We call the vector c R n having coordinates c1, c2 . . . cn the “coordinate vector” of x with
respect to the basis B, and denote it by [x]B. 1.3.
First examples. (a) [x]E = x for each x R n . (b) For any basis B = {v1, v2, . . . vn} of R n : [vj
]B = ej (1 j n). (c) [0]B = 0 for any basis B (the zero vector has the same representation with
respect to any basis.
1.4. Matrix calculation of coordinates. Letting B = {v1, v2, . . . vn} be a basis for R n we can
rewrite equation (1.1) in matrix form as (1.2) x = V [x]B where V is the matrix having the vector
vj for its j-th column, i.e. V = [v1, v2, . . . vn].
. 1 2 EIGENVALUES/EIGENVECTORS V is invertible because its columns are linearly
independent
(see, for example, Summary 3.2.9(v) of our textbook, pp. 119–120), so we can compute the B
coordinate representation of any vector x R n by solving (1.2) for [x]B: (1.3) [x]B = V 1x For
example, if in R 2 we take B = {v1, v2} to be our basis, where v1 = 1 3 and v2 = 2 5 , then V = 1
2 3 5 and V 1 = 5 2 3 1 .
Suppose, for instance, that x = 1 1 . Then [x]B = V 1x = 5 2 3 1 1 1 = 7 4 More generally, if x
= x1 x2 is any vector in R n then the same calculation shows that [x]B = 5x1 + 2x2 3x1 x2 .
1.5. The matrix of a linear transformation with respect to a basis. Recall that each n × n matrix
induces a linear transformation T : R n R n by means of matrix multiplication: (1.4) T(x) = Ax
(x R n ). We refer to A as the “standard matrix” for T. The lesson of what’s to follow is that it’s
not always the best matrix for T! For a basis B = {v1, v2, . . . vn} for R n we define (1.5) [T]B =
[[T(v1)]B, [T(v2)]B . . . [T(vn)]B] , i.e. [T]B is the matrix that has the B-coordinate
representation of T(vj ) as its j-th column (1 j n). We call [T]B the matrix of T with respect to
the basis B.
1.6. Example: B = E, the standard basis. In this case [T]B = [Ae1, Ae2, . . . Aen], the “standard
matrix” for T. EIGENVALUES/EIGENVECTORS 3 1.7. Geometric examples in R 2 . In this
next group of examples, L is a line through the origin in the direction of a non-zero vector v R 2
, and w is a non-zero vector perpendicular to v. We take as our basis for R 2 , not the standard
one, but instead B = {v, w}.
(a) Suppose T : R 2 R 2 is the transformation of orthogonal projection onto L. Then T(v) = v and
T(w) = 0, so [T]B = [[T(v)]B, [T(w)]B] = [[v]B, [0]B] = [e1, 0] = 1 0 0 0 , where we use the fact
that [v]B = e1 and [0]B = 0 (see §1.3 (b) and (c)). Compare this simple and informative matrix
for T with the one we got when we first introduced orthogonal projection. Then we took v to be a
unit vector with coordinates u1 and u2, and did this computation:
T(x) = (x · v)v = (x1u2 + x2u2) u1 u2 = u 2 1x1 + u1u2x2 u1u2x1 + u 2 2x2 = u 2 1 u1u2 u1u2 u
2 2 x1 x2 , so the “standard matrix” for the linear transformation of orthogonal projection onto L
is the completely uninformative one A = u 2 1 u1u2 u1u2 u 2 2
. (b) Suppose T is reflection in L. Then T(v) = v as before, but now T(w) = w. Thus [T]B =
[[T(v)]B, [T(w)]B] = [[v]B, [w]B] = [e1, e2] = 1 0 0 1 .
Note that in this computation we used the fact that [w]B = [w]B (“linearity of
coordinates”—Fact 3.4.2, page 139 of our textbook) and that [v]B = e1 and [w]B = e2 (§1.3 (b)).
(c) Suppose T is a shear in the direction of L, i.e. that T(v) = v and T(w) = kv + w for some fixed
real number k. Then using linearity of coordinates and §1.3(b) we see that [T]B = 1 k 0 1 which
is the same as the standard matrix for a horizontal shear.
4 EIGENVALUES/EIGENVECTORS
1.8. Using [T]B to compute T(x). The utility of the matrix of a linear transformation T with
respect to a basis B lies in the following formula:
(1.6) [T]B[x]B = [T(x)]B (x R n ), which says that, if you’re willing to give up the standard
basis fro R n , and do all your calculations relative to the basis B, then you can use the matrix of
T with respect to B in the same way you used to use the standard matrix.
Proof of formula (1.6): For x R n we have the unique representation x = c1v1 + c2v2 + . . . +
cnvn where c1, c2, . . . , cn are the (standard) coordinates of [x]B.
By linearity of T: T(x) = c1T(v1) + c2T(v2) + . . . + cnT(vn), so by linearity of coordinates:
[T(x)]B = c1[T(v1)]B + c2[T(v2)]B + . . . + cn[T(vn)]B,
But the expression on the right is just the matrix product [T]B[x]B, so we are done.
2. Eigenvalues and Eigenvectors
2.1. Definition. Suppose A is an n×n matrix and T the associated linear transformation of R n (so
T(x) = Ax for every x R n ).
We call a real number an eigenvalue of T (or of A) whenever there is a non-zero vector x R n
such that T(x) = x.
The vector x is called an eigenvector of T (or of A) corresponding to the eigenvalue .
2.2. Examples. (a) Look back at the examples of §1.7, where L is a line through the origin in the
direction of a non-zero vector v, and w is a non-zero vector perpendicular to v.
If T : R 2 R 2 is orthogonal projection onto L, then v and w are eigenvectors of T corresponding
respectively to eigenvalues 1 and 0.
If T is reflection in L then v and w are still eigenvectors, but now the corresponding eigenvalues
are 1 and 1.
If T is a shear along L then v is an eigenvector with eigenvalue 1, but w is not an eigenvector.
EIGENVALUES/EIGENVECTORS
5 (b) (see Problem 25, page 146). Suppose A = 1 2 4 3 and v = 1 2 , w = 1 1 .
Then one computes that Av = 5v and Aw = w, so v is an eigenvector for A corresponding to the
eigenvalue 5 and w is an eigenvector corresponding to the eigenvalue 1.
Note that, with respect to the basis B = {v, w} of R 2 , the matrix for the linear transformation
induced by A is just 5 0 0 1 .
(c) (see Problem 29 of §3.4). Here T is the linear transformation of R 3 induced by the matrix A
= 1 1 0 0 2 2 3 9 6 We are given a basis B = {v1, v2, v3} for R 3 , where v1 = 1 1 1 , v2 = 1 2 3
, and v3 = 1 3 6
. You can check easily that each of these basis vectors is an eigenvector: Av1 = 0 (eigenvalue
zero), Av2 = v2 (eigenvalue 1) and Av3 = 2v3 (eigenvalue 2), and that therefore the matrix of
the associated linear transformation T with respect to the basis B is [T]B = 0 0 0 0 1 0 0 0 2
In each of the examples of the previous two sections, whenever we had a linear transformation T
of R n that bequeathed to R n a basis B of eigenvectors, the matrix of T with respect to B turned
out to be diagonal, i.e. all entries zero off the main diagonal. 2.3. Notation. The n × n diagonal
matrix with entry j in the j-th diagonal position will be denoted diag {1, 2, . . . n}. Thus, for
example, the 3 × 3 identity matrix is diag {1, 1, 1}, while in §2.2(c), [T]B = diag {0, 1, 2}.
Diagonal matrices are the easiest ones for which to find eigenvalues and eigenvectors:
2.4. Proposition. If A = diag {1, 2, . . . n} then the standard unit vector ej is an eigenvector for
the eigenvalue j (1 j n).
6 EIGENVALUES/EIGENVECTORS Proof. Our hypothesis is that A = [1e1, 2e2, . . . , nen],
i.e. that A is the matrix with jej as its j-th column (1 j n).
But Aej is the j-th column of A, so Aej = jej for each j between 1 and n.
2.5. Exercise. Prove the converse: If an n × n matrix A has the standard basis for R n as
eigenvectors, then A is diagonal.
2.6. The Diagonalization Theorem. For a linear transformation T : R n R n , and a basis B for R
n , the following two statements are equivalent (each implies the other):
(a) B consists of eigenvectors of T. (b) [T]B is a diagonal matrix.
Proof. (a) (b): Suppose B = {v1, v2, . . . vn} consists of eigenvectors of T, so that there are real
numbers j (1 j n) such that T(vj ) = jvj (j = 1, 2, . . . , n).
We must show that [T]B is diagonal. Starting with the definition of [T]B, we compute: [T]B =
[[T v1]B, [T v2]B, . . . , [T vn]B] = [[1v1]B, [2v2]B, . . . , [nvn]B] = [1[v1]B, 2[v2]B, . . . ,
n[vn]B] = [1e1, 2e2, . . . , nen] = diag {1, 2, . . . , n} where, in this computation we used
successively: the definition of [T]B (line 1), the fact that T(vj ) = jvj (line 2), linearity of
coordinates (line 3), and finally the fact that [vj ]B = ej (line 4).
(b) (a): Now suppose B = {v1, v2, . . . vn} is a basis for R n , and that [T]B = diag {1, 2, . . . , n}.
We must show that each vj is an eigenvector of T. In fact, we’ll see that T vj = jvj for 1 j n. Fix
such a j, and note that the j-th column of [T]B is just jej , i.e.,
(2.1) [T]Bej = jej EIGENVALUES/EIGENVECTORS 7 Now the right-hand side of (2.1) is j [vj
]B which, by linearity of coordinates, is [jvj ]B. Similarly, the left-hand side of (2.1) is [T]B[vj
]B which, by (1.6), is [T vj ]B. Thus [T vj ]B = [jvj ]B, i.e., the both T vj and jvj have the same
B-coordinate vectors.
Thus T vj = jvj , as desired. 2.7. Definition. Call an n × n matrix diagonalizable if R n has a basis
consisting of eigenvectors of this matrix. Warning: Not all square matrices are diagonalizable!
For example it’s geometrically clear that shear transfomations of R 2 have (up to multiplication
by a non-zero constant) just a single eigenvector, while rotations of R 2 through any angle not an
integer multiple of have no eigenvectors at all! We’ll see how to prove this kind of
result“analytically” (i.e. without reference to pictures) in the next section.
3. Efficient eigenvalue computation for 2 × 2 matrices We’ll base our computations for 2 × 2
matrices on the notion of determinant.
3.1. Definition. The determinant of the matrix A = a b c d is: det(A) = ad bc The key to all the
work of this section is: 3.2. Theorem. For A a 2 × 2 matrix, the system of linear equations Ax = 0
has nontrivial solution if and only if det(A) = 0.
Proof. In scalar form the system of equations is: ax1 + bx2 = 0 cx1 + dx2 = 0 Suppose first that
neither a nor c is zero. Then we can replace the second equation by c times the first minus a
times the second, yielding the equivalent system: ax1 + bx2 = 0 (ad bc)x2 = 0 8
EIGENVALUES/EIGENVECTORS which has a nontrivial solution if and only if ad bc = 0.
Suppose a = 0. Then det(A) = bc, which is zero precisely when either b or c is zero. If b = 0 then
we are back to one equation in two unknowns, so there are nontrivial (infinitely many) solutions.
If b 6= 0 then any solution must have x2 = 0, whereupon the second equation implies cx1 = 0. If
c = 0 (i.e. det(A) = 0) then there are infinitely many solutions: x2 = 0 and x1 arbitrary. If,
however c 6= 0 (i.e., det(A) 6= 0) then x1 is also required to be zero, hence the only solution is
the trivial one.
3.3. Corollary. The eigenvalues of a 2 × 2 matrix A are the real numbers that satisfy its
“characteristic equation” det(a I) = 0.
Proof. This follows immediately from the above Theorem and the definition of eigenvalue. More
precisely, to say a real number is an eigenvalue of A means that there is a non-zero vector x R 2
such that: Ax = x i.e. Ax x = 0 i.e. (A I)x = 0 Thus is an eigenvector of A if and only if the
homogeneous system (A I)x = 0 has a nontrivial solution, which, by Theorem 3.2, happens if
and only if det(A I) = 0. 3.4.
Example. Show that the matrix A = 1 2 4 3 is diagonalizable.
Proof. Of course we know this already from §2.2(b) where the eigenvectors and eigenvalues of A
were handed to us. But suppose they weren’t. Then we write the characteristic equation: 0 =
det(A I) = det( 1 2 4 3 ) = (1 )(3 ) 8 = 2 4 5 = ( 5)( + 1)
EIGENVALUES/EIGENVECTORS 9 so that = 5 and = 1 are the solutions, and so these are the
eigenvalues of A. Now for = 5, A I = A 5I = 4 2 4 2 and the system (A 5I)x = 0 is easily seen
to have solutions x = any scalar multiple of 1 2 . In particular, v1 = 1 2 is the unique (up to
multiplication by a non-zero scalar) eigenvector for the eigenvalue 5. I leave it to you to show by
the same method that v2 = 1 1 is the “unique” eigenvector for the eigenvalue 1. It’s easy to
check that B = {v1, v2} is a basis for R 2 , so A is diagonalizable, and the matrix of its linear
transformation, with respect to the basis B is diag {5, 1}. 3.5. Example: a non-diagonalizable
matrix. Let A = 0 1 0 0 . The characteristic equation for this matrix is 2 = 0, which has only = 0
as its solution. The corresponding eigenvectors are the nontrivial solutions to the equation Ax =
0 (i.e. the non-zero vectors in the kernel of A), and these all the non-zero scalar multiples of the
standard unit vector e1. Thus there aren’t enough eigenvectors here to make a basis of R 2 , so A
is not diagonalizable. 4. Matrix of [T]B vs. standard matrix of T Suppose T is a linear
transformation of R n whows “standard matrix” is A, i.e. T(x) = Ax (x R n ). Suppose B = {v1,
v2, . . . vn} is a basis for R n . It’s time to get serious about the relationship between A and [T]B.
Recall that in §1.8 we derived equation (1.3): [x]B = V 1x where V = [v1, v2, . . . , vn] is the
matrix having the vector vj as its j-th column. Recall also that the invertibility of V derives from
the fact that its columns are linearly independent.
10 EIGENVALUES/EIGENVECTORS
Thus we may compute: [T]B = [[T(v1)]B, [T(v2)]B, . . . , [T(vn)]B] = [[Av1]B, [Av2]B, . . . ,
[Avn]B] = V 1Av1, V 1Av2, . . . , V 1Avn = V 1A [v1, v2, . . . , vn] = V 1AV where the first line
is the definition of [T]B, the second uses the definition of T in terms of A, the third uses (1.3),
and the fourth is the definition of matrix multiplication. We have proved: 4.1. Theorem (Fact
3.4.4, Page 144 of our text). Suppose B = {v1, v2, . . . vn} is a basis for R n and T a linear
transformation of R n with standard matrix A. Then [T]B = V 1AV , where V is the matrix
whose j-th column is vj (1 j n). 5. Application: Powers of matrices Theorem 4.1 is of great
utility in computing large powers of matrices—a calculation that would be difficult to do by
hand in many circumstances. The method depends on two crucial facts. 5.1. Theorem. If A, B
and V are n × n matrices, and V is invertible, then for each positive integer p: (V BV 1 ) p = V
BpV 1 . Proof. We use the method of Mathematical Induction, which asserts that to prove the
theorem we need only check that it’s true for p = 1, and then show that: (*) Result true for an
integer p 1 Result true for p + 1. The point is that once you’ve established the case p = 1 and
proved (*), then (*) gives you for free the case p = 2, and then the case p = 3, and so on forever.
EIGENVALUES/EIGENVECTORS 11 For our result, the case p = 1 is obviously true, so
suppose we have proved the theorem for some p 1. We must show that it’s true for p + 1; to this
end write (V BV 1 ) p+1 = (V BV 1 ) p (V BV 1 ) = (V BpV 1 )(V BV 1 ) = (V Bp )(V 1V )(BV
1 ) = (V Bp )I(BV 1 ) = V Bp+1V 1 where the second line follows from our “induction
hypothesis” that the theorem is true for p, the third line follows from the associate law of matrix
multiplication, and the others are clear. The next result asserts that for a diagonal matrix you can
compute a power by taking the corresponding power of each element of the matrix. 5.2.
Theorem. For any positive integer p, diag {1, 2, . . . , n} p = diag { p 1 , p 2 , . . . , p n} Proof. I
leave this one mostly as an exercise. First check that diag {1, 2, . . . , n}diag {µ1, µ2, . . . , µn} =
diag {1µ1, 2µ2, . . . , nµn} The result about powers then follows quickly from this by induction.
5.3. Example. Compute 1 2 4 3 15 .

More Related Content

Similar to Hi please complete the following with detailed working out Find the .pdf

linear transformation and rank nullity theorem
linear transformation and rank nullity theorem linear transformation and rank nullity theorem
linear transformation and rank nullity theorem Manthan Chavda
 
Analytical Geometry in three dimension
Analytical Geometry in three dimensionAnalytical Geometry in three dimension
Analytical Geometry in three dimensionSwathiSundari
 
Notes on eigenvalues
Notes on eigenvaluesNotes on eigenvalues
Notes on eigenvaluesAmanSaeed11
 
Vcla ppt ch=vector space
Vcla ppt ch=vector spaceVcla ppt ch=vector space
Vcla ppt ch=vector spaceMahendra Patel
 
On the lambert w function
On the lambert w functionOn the lambert w function
On the lambert w functionTrungKienVu3
 
Mat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptMat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptabidraufv
 
The concept of moments of order invariant quantum
The concept of moments of order invariant quantumThe concept of moments of order invariant quantum
The concept of moments of order invariant quantumAlexander Decker
 
Reviewof vectoranalysis
Reviewof vectoranalysisReviewof vectoranalysis
Reviewof vectoranalysisArun Soni
 
2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and DimensionCeni Babaoglu, PhD
 
Interpolation functions for linear triangle elements (very elementary)
Interpolation functions for linear triangle elements (very elementary)Interpolation functions for linear triangle elements (very elementary)
Interpolation functions for linear triangle elements (very elementary)Prashant K. Jha
 
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear TransformationsCeni Babaoglu, PhD
 
Linear Algebra Gauss Jordan elimination.pptx
Linear Algebra Gauss Jordan elimination.pptxLinear Algebra Gauss Jordan elimination.pptx
Linear Algebra Gauss Jordan elimination.pptxMaths Assignment Help
 
Understanding viscoelasticity
Understanding viscoelasticityUnderstanding viscoelasticity
Understanding viscoelasticitySpringer
 
We now prove for every finite-dimensional complex vector space V an.pdf
We now prove for every finite-dimensional complex vector space V an.pdfWe now prove for every finite-dimensional complex vector space V an.pdf
We now prove for every finite-dimensional complex vector space V an.pdfdavid16271
 

Similar to Hi please complete the following with detailed working out Find the .pdf (20)

linear transformation and rank nullity theorem
linear transformation and rank nullity theorem linear transformation and rank nullity theorem
linear transformation and rank nullity theorem
 
Analytical Geometry in three dimension
Analytical Geometry in three dimensionAnalytical Geometry in three dimension
Analytical Geometry in three dimension
 
Notes on eigenvalues
Notes on eigenvaluesNotes on eigenvalues
Notes on eigenvalues
 
Vcla ppt ch=vector space
Vcla ppt ch=vector spaceVcla ppt ch=vector space
Vcla ppt ch=vector space
 
Vectouurs
VectouursVectouurs
Vectouurs
 
Vectouurs
VectouursVectouurs
Vectouurs
 
On the lambert w function
On the lambert w functionOn the lambert w function
On the lambert w function
 
Signals Processing Homework Help
Signals Processing Homework HelpSignals Processing Homework Help
Signals Processing Homework Help
 
Mat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.pptMat 223_Ch4-VectorSpaces.ppt
Mat 223_Ch4-VectorSpaces.ppt
 
Curves in space
Curves in spaceCurves in space
Curves in space
 
Linear algebra
Linear algebraLinear algebra
Linear algebra
 
The concept of moments of order invariant quantum
The concept of moments of order invariant quantumThe concept of moments of order invariant quantum
The concept of moments of order invariant quantum
 
Reviewof vectoranalysis
Reviewof vectoranalysisReviewof vectoranalysis
Reviewof vectoranalysis
 
2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension2. Linear Algebra for Machine Learning: Basis and Dimension
2. Linear Algebra for Machine Learning: Basis and Dimension
 
Interpolation functions for linear triangle elements (very elementary)
Interpolation functions for linear triangle elements (very elementary)Interpolation functions for linear triangle elements (very elementary)
Interpolation functions for linear triangle elements (very elementary)
 
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
3. Linear Algebra for Machine Learning: Factorization and Linear Transformations
 
150490106037
150490106037150490106037
150490106037
 
Linear Algebra Gauss Jordan elimination.pptx
Linear Algebra Gauss Jordan elimination.pptxLinear Algebra Gauss Jordan elimination.pptx
Linear Algebra Gauss Jordan elimination.pptx
 
Understanding viscoelasticity
Understanding viscoelasticityUnderstanding viscoelasticity
Understanding viscoelasticity
 
We now prove for every finite-dimensional complex vector space V an.pdf
We now prove for every finite-dimensional complex vector space V an.pdfWe now prove for every finite-dimensional complex vector space V an.pdf
We now prove for every finite-dimensional complex vector space V an.pdf
 

More from ezhilvizhiyan

Explain and discuss 4G wireless communications and their advantages..pdf
Explain and discuss 4G wireless communications and their advantages..pdfExplain and discuss 4G wireless communications and their advantages..pdf
Explain and discuss 4G wireless communications and their advantages..pdfezhilvizhiyan
 
Draw and describe a module of the cross-cultural communication pr.pdf
Draw and describe a module of the cross-cultural communication pr.pdfDraw and describe a module of the cross-cultural communication pr.pdf
Draw and describe a module of the cross-cultural communication pr.pdfezhilvizhiyan
 
Discuss the reasons why visions fail. What steps can be implemented .pdf
Discuss the reasons why visions fail. What steps can be implemented .pdfDiscuss the reasons why visions fail. What steps can be implemented .pdf
Discuss the reasons why visions fail. What steps can be implemented .pdfezhilvizhiyan
 
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdf
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdfDescribe the original purpose of the Clean Air Act Policy of 1963. E.pdf
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdfezhilvizhiyan
 
Continuity 100 Let f be a continuous function on a metric space X. L.pdf
Continuity 100 Let f be a continuous function on a metric space X. L.pdfContinuity 100 Let f be a continuous function on a metric space X. L.pdf
Continuity 100 Let f be a continuous function on a metric space X. L.pdfezhilvizhiyan
 
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdf
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdfAdmitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdf
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdfezhilvizhiyan
 
7. In many respects metal binding is similar to the binding of a prot.pdf
7. In many respects metal binding is similar to the binding of a prot.pdf7. In many respects metal binding is similar to the binding of a prot.pdf
7. In many respects metal binding is similar to the binding of a prot.pdfezhilvizhiyan
 
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdfezhilvizhiyan
 
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdfezhilvizhiyan
 
X Company has the following budgeted cash flows for JanuaryIf the .pdf
X Company has the following budgeted cash flows for JanuaryIf the .pdfX Company has the following budgeted cash flows for JanuaryIf the .pdf
X Company has the following budgeted cash flows for JanuaryIf the .pdfezhilvizhiyan
 
Why did animal phyla appear so suddenly during the Cambrian explosion.pdf
Why did animal phyla appear so suddenly during the Cambrian explosion.pdfWhy did animal phyla appear so suddenly during the Cambrian explosion.pdf
Why did animal phyla appear so suddenly during the Cambrian explosion.pdfezhilvizhiyan
 
Which of the following information-management systems uses artificia.pdf
Which of the following information-management systems uses artificia.pdfWhich of the following information-management systems uses artificia.pdf
Which of the following information-management systems uses artificia.pdfezhilvizhiyan
 
Which of the following was part of the reason for the European excha.pdf
Which of the following was part of the reason for the European excha.pdfWhich of the following was part of the reason for the European excha.pdf
Which of the following was part of the reason for the European excha.pdfezhilvizhiyan
 
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdf
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdfTure or false or uncertain ( explain why) Thank you! e. Output per c.pdf
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdfezhilvizhiyan
 
This is for a C programDene a Car structure type in your header le.pdf
This is for a C programDene a Car structure type in your header le.pdfThis is for a C programDene a Car structure type in your header le.pdf
This is for a C programDene a Car structure type in your header le.pdfezhilvizhiyan
 
This project will implement a simple usernamepassword lookup system.pdf
This project will implement a simple usernamepassword lookup system.pdfThis project will implement a simple usernamepassword lookup system.pdf
This project will implement a simple usernamepassword lookup system.pdfezhilvizhiyan
 
The following code is based on the Josephus problem, the code does c.pdf
The following code is based on the Josephus problem, the code does c.pdfThe following code is based on the Josephus problem, the code does c.pdf
The following code is based on the Josephus problem, the code does c.pdfezhilvizhiyan
 
Tech transfers should the fed Gov’t keep subsidizing University .pdf
Tech transfers should the fed Gov’t keep subsidizing University .pdfTech transfers should the fed Gov’t keep subsidizing University .pdf
Tech transfers should the fed Gov’t keep subsidizing University .pdfezhilvizhiyan
 
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdf
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdfRemaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdf
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdfezhilvizhiyan
 
QUESTION 9 Research on North American women serving as expatriates fo.pdf
QUESTION 9 Research on North American women serving as expatriates fo.pdfQUESTION 9 Research on North American women serving as expatriates fo.pdf
QUESTION 9 Research on North American women serving as expatriates fo.pdfezhilvizhiyan
 

More from ezhilvizhiyan (20)

Explain and discuss 4G wireless communications and their advantages..pdf
Explain and discuss 4G wireless communications and their advantages..pdfExplain and discuss 4G wireless communications and their advantages..pdf
Explain and discuss 4G wireless communications and their advantages..pdf
 
Draw and describe a module of the cross-cultural communication pr.pdf
Draw and describe a module of the cross-cultural communication pr.pdfDraw and describe a module of the cross-cultural communication pr.pdf
Draw and describe a module of the cross-cultural communication pr.pdf
 
Discuss the reasons why visions fail. What steps can be implemented .pdf
Discuss the reasons why visions fail. What steps can be implemented .pdfDiscuss the reasons why visions fail. What steps can be implemented .pdf
Discuss the reasons why visions fail. What steps can be implemented .pdf
 
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdf
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdfDescribe the original purpose of the Clean Air Act Policy of 1963. E.pdf
Describe the original purpose of the Clean Air Act Policy of 1963. E.pdf
 
Continuity 100 Let f be a continuous function on a metric space X. L.pdf
Continuity 100 Let f be a continuous function on a metric space X. L.pdfContinuity 100 Let f be a continuous function on a metric space X. L.pdf
Continuity 100 Let f be a continuous function on a metric space X. L.pdf
 
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdf
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdfAdmitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdf
Admitting New Partner With Bonus Cody Jenkins and Lacey Tanner formed.pdf
 
7. In many respects metal binding is similar to the binding of a prot.pdf
7. In many respects metal binding is similar to the binding of a prot.pdf7. In many respects metal binding is similar to the binding of a prot.pdf
7. In many respects metal binding is similar to the binding of a prot.pdf
 
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf
3. Photosynthetic organisms produce about 300 x 1015 g of oxygen per .pdf
 
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf
1. Which is less soluble in water, 1-pentanol or 1-heptanol Explain..pdf
 
X Company has the following budgeted cash flows for JanuaryIf the .pdf
X Company has the following budgeted cash flows for JanuaryIf the .pdfX Company has the following budgeted cash flows for JanuaryIf the .pdf
X Company has the following budgeted cash flows for JanuaryIf the .pdf
 
Why did animal phyla appear so suddenly during the Cambrian explosion.pdf
Why did animal phyla appear so suddenly during the Cambrian explosion.pdfWhy did animal phyla appear so suddenly during the Cambrian explosion.pdf
Why did animal phyla appear so suddenly during the Cambrian explosion.pdf
 
Which of the following information-management systems uses artificia.pdf
Which of the following information-management systems uses artificia.pdfWhich of the following information-management systems uses artificia.pdf
Which of the following information-management systems uses artificia.pdf
 
Which of the following was part of the reason for the European excha.pdf
Which of the following was part of the reason for the European excha.pdfWhich of the following was part of the reason for the European excha.pdf
Which of the following was part of the reason for the European excha.pdf
 
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdf
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdfTure or false or uncertain ( explain why) Thank you! e. Output per c.pdf
Ture or false or uncertain ( explain why) Thank you! e. Output per c.pdf
 
This is for a C programDene a Car structure type in your header le.pdf
This is for a C programDene a Car structure type in your header le.pdfThis is for a C programDene a Car structure type in your header le.pdf
This is for a C programDene a Car structure type in your header le.pdf
 
This project will implement a simple usernamepassword lookup system.pdf
This project will implement a simple usernamepassword lookup system.pdfThis project will implement a simple usernamepassword lookup system.pdf
This project will implement a simple usernamepassword lookup system.pdf
 
The following code is based on the Josephus problem, the code does c.pdf
The following code is based on the Josephus problem, the code does c.pdfThe following code is based on the Josephus problem, the code does c.pdf
The following code is based on the Josephus problem, the code does c.pdf
 
Tech transfers should the fed Gov’t keep subsidizing University .pdf
Tech transfers should the fed Gov’t keep subsidizing University .pdfTech transfers should the fed Gov’t keep subsidizing University .pdf
Tech transfers should the fed Gov’t keep subsidizing University .pdf
 
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdf
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdfRemaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdf
Remaining Time 31 minutes, 22 seconds QUESTION 1 Question Completion.pdf
 
QUESTION 9 Research on North American women serving as expatriates fo.pdf
QUESTION 9 Research on North American women serving as expatriates fo.pdfQUESTION 9 Research on North American women serving as expatriates fo.pdf
QUESTION 9 Research on North American women serving as expatriates fo.pdf
 

Recently uploaded

Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfSumit Tiwari
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceSamikshaHamane
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxiammrhaywood
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerunnathinaik
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaVirag Sontakke
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon AUnboundStockton
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxEyham Joco
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfMahmoud M. Sallam
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...M56BOOKSTORE PRODUCT/SERVICE
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 

Recently uploaded (20)

Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in Pharmacovigilance
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developer
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of India
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon A
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptx
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
OS-operating systems- ch04 (Threads) ...
OS-operating systems- ch04 (Threads) ...OS-operating systems- ch04 (Threads) ...
OS-operating systems- ch04 (Threads) ...
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdf
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
 
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 

Hi please complete the following with detailed working out Find the .pdf

  • 1. Hi please complete the following with detailed working out Find the eigenvalues of C1 C1 C1 C1 C2 C3 C3 C3 C3 cn cn Hint: You may find that the identity A2 (c1 c2 cm) is useful. Solution COORDINATES, EIGENVALUES, AND EIGENVECTORS 1. Review of coordinates 1.1. The standard basis. The standard basis for R n is the basis E = {e1, e2, . . . en}, where ej is the vector with 1 in the j-th position and zeros elsewhere. 1.2. Coordinates with respect to a basis. Let B = {v1, v2, . . . vn} be a basis for R n . We know that for each vector x R n there are unique scalar “coordinates” c1, c2, . . . cn such that (1.1) x = c1v1 + c2v2 + . . . + cnvn We call the vector c R n having coordinates c1, c2 . . . cn the “coordinate vector” of x with respect to the basis B, and denote it by [x]B. 1.3. First examples. (a) [x]E = x for each x R n . (b) For any basis B = {v1, v2, . . . vn} of R n : [vj ]B = ej (1 j n). (c) [0]B = 0 for any basis B (the zero vector has the same representation with respect to any basis. 1.4. Matrix calculation of coordinates. Letting B = {v1, v2, . . . vn} be a basis for R n we can rewrite equation (1.1) in matrix form as (1.2) x = V [x]B where V is the matrix having the vector vj for its j-th column, i.e. V = [v1, v2, . . . vn]. . 1 2 EIGENVALUES/EIGENVECTORS V is invertible because its columns are linearly independent (see, for example, Summary 3.2.9(v) of our textbook, pp. 119–120), so we can compute the B coordinate representation of any vector x R n by solving (1.2) for [x]B: (1.3) [x]B = V 1x For example, if in R 2 we take B = {v1, v2} to be our basis, where v1 = 1 3 and v2 = 2 5 , then V = 1 2 3 5 and V 1 = 5 2 3 1 . Suppose, for instance, that x = 1 1 . Then [x]B = V 1x = 5 2 3 1 1 1 = 7 4 More generally, if x = x1 x2 is any vector in R n then the same calculation shows that [x]B = 5x1 + 2x2 3x1 x2 . 1.5. The matrix of a linear transformation with respect to a basis. Recall that each n × n matrix induces a linear transformation T : R n R n by means of matrix multiplication: (1.4) T(x) = Ax (x R n ). We refer to A as the “standard matrix” for T. The lesson of what’s to follow is that it’s not always the best matrix for T! For a basis B = {v1, v2, . . . vn} for R n we define (1.5) [T]B = [[T(v1)]B, [T(v2)]B . . . [T(vn)]B] , i.e. [T]B is the matrix that has the B-coordinate representation of T(vj ) as its j-th column (1 j n). We call [T]B the matrix of T with respect to the basis B.
  • 2. 1.6. Example: B = E, the standard basis. In this case [T]B = [Ae1, Ae2, . . . Aen], the “standard matrix” for T. EIGENVALUES/EIGENVECTORS 3 1.7. Geometric examples in R 2 . In this next group of examples, L is a line through the origin in the direction of a non-zero vector v R 2 , and w is a non-zero vector perpendicular to v. We take as our basis for R 2 , not the standard one, but instead B = {v, w}. (a) Suppose T : R 2 R 2 is the transformation of orthogonal projection onto L. Then T(v) = v and T(w) = 0, so [T]B = [[T(v)]B, [T(w)]B] = [[v]B, [0]B] = [e1, 0] = 1 0 0 0 , where we use the fact that [v]B = e1 and [0]B = 0 (see §1.3 (b) and (c)). Compare this simple and informative matrix for T with the one we got when we first introduced orthogonal projection. Then we took v to be a unit vector with coordinates u1 and u2, and did this computation: T(x) = (x · v)v = (x1u2 + x2u2) u1 u2 = u 2 1x1 + u1u2x2 u1u2x1 + u 2 2x2 = u 2 1 u1u2 u1u2 u 2 2 x1 x2 , so the “standard matrix” for the linear transformation of orthogonal projection onto L is the completely uninformative one A = u 2 1 u1u2 u1u2 u 2 2 . (b) Suppose T is reflection in L. Then T(v) = v as before, but now T(w) = w. Thus [T]B = [[T(v)]B, [T(w)]B] = [[v]B, [w]B] = [e1, e2] = 1 0 0 1 . Note that in this computation we used the fact that [w]B = [w]B (“linearity of coordinates”—Fact 3.4.2, page 139 of our textbook) and that [v]B = e1 and [w]B = e2 (§1.3 (b)). (c) Suppose T is a shear in the direction of L, i.e. that T(v) = v and T(w) = kv + w for some fixed real number k. Then using linearity of coordinates and §1.3(b) we see that [T]B = 1 k 0 1 which is the same as the standard matrix for a horizontal shear. 4 EIGENVALUES/EIGENVECTORS 1.8. Using [T]B to compute T(x). The utility of the matrix of a linear transformation T with respect to a basis B lies in the following formula: (1.6) [T]B[x]B = [T(x)]B (x R n ), which says that, if you’re willing to give up the standard basis fro R n , and do all your calculations relative to the basis B, then you can use the matrix of T with respect to B in the same way you used to use the standard matrix. Proof of formula (1.6): For x R n we have the unique representation x = c1v1 + c2v2 + . . . + cnvn where c1, c2, . . . , cn are the (standard) coordinates of [x]B. By linearity of T: T(x) = c1T(v1) + c2T(v2) + . . . + cnT(vn), so by linearity of coordinates: [T(x)]B = c1[T(v1)]B + c2[T(v2)]B + . . . + cn[T(vn)]B, But the expression on the right is just the matrix product [T]B[x]B, so we are done. 2. Eigenvalues and Eigenvectors 2.1. Definition. Suppose A is an n×n matrix and T the associated linear transformation of R n (so T(x) = Ax for every x R n ). We call a real number an eigenvalue of T (or of A) whenever there is a non-zero vector x R n such that T(x) = x.
  • 3. The vector x is called an eigenvector of T (or of A) corresponding to the eigenvalue . 2.2. Examples. (a) Look back at the examples of §1.7, where L is a line through the origin in the direction of a non-zero vector v, and w is a non-zero vector perpendicular to v. If T : R 2 R 2 is orthogonal projection onto L, then v and w are eigenvectors of T corresponding respectively to eigenvalues 1 and 0. If T is reflection in L then v and w are still eigenvectors, but now the corresponding eigenvalues are 1 and 1. If T is a shear along L then v is an eigenvector with eigenvalue 1, but w is not an eigenvector. EIGENVALUES/EIGENVECTORS 5 (b) (see Problem 25, page 146). Suppose A = 1 2 4 3 and v = 1 2 , w = 1 1 . Then one computes that Av = 5v and Aw = w, so v is an eigenvector for A corresponding to the eigenvalue 5 and w is an eigenvector corresponding to the eigenvalue 1. Note that, with respect to the basis B = {v, w} of R 2 , the matrix for the linear transformation induced by A is just 5 0 0 1 . (c) (see Problem 29 of §3.4). Here T is the linear transformation of R 3 induced by the matrix A = 1 1 0 0 2 2 3 9 6 We are given a basis B = {v1, v2, v3} for R 3 , where v1 = 1 1 1 , v2 = 1 2 3 , and v3 = 1 3 6 . You can check easily that each of these basis vectors is an eigenvector: Av1 = 0 (eigenvalue zero), Av2 = v2 (eigenvalue 1) and Av3 = 2v3 (eigenvalue 2), and that therefore the matrix of the associated linear transformation T with respect to the basis B is [T]B = 0 0 0 0 1 0 0 0 2 In each of the examples of the previous two sections, whenever we had a linear transformation T of R n that bequeathed to R n a basis B of eigenvectors, the matrix of T with respect to B turned out to be diagonal, i.e. all entries zero off the main diagonal. 2.3. Notation. The n × n diagonal matrix with entry j in the j-th diagonal position will be denoted diag {1, 2, . . . n}. Thus, for example, the 3 × 3 identity matrix is diag {1, 1, 1}, while in §2.2(c), [T]B = diag {0, 1, 2}. Diagonal matrices are the easiest ones for which to find eigenvalues and eigenvectors: 2.4. Proposition. If A = diag {1, 2, . . . n} then the standard unit vector ej is an eigenvector for the eigenvalue j (1 j n). 6 EIGENVALUES/EIGENVECTORS Proof. Our hypothesis is that A = [1e1, 2e2, . . . , nen], i.e. that A is the matrix with jej as its j-th column (1 j n). But Aej is the j-th column of A, so Aej = jej for each j between 1 and n. 2.5. Exercise. Prove the converse: If an n × n matrix A has the standard basis for R n as eigenvectors, then A is diagonal. 2.6. The Diagonalization Theorem. For a linear transformation T : R n R n , and a basis B for R n , the following two statements are equivalent (each implies the other): (a) B consists of eigenvectors of T. (b) [T]B is a diagonal matrix.
  • 4. Proof. (a) (b): Suppose B = {v1, v2, . . . vn} consists of eigenvectors of T, so that there are real numbers j (1 j n) such that T(vj ) = jvj (j = 1, 2, . . . , n). We must show that [T]B is diagonal. Starting with the definition of [T]B, we compute: [T]B = [[T v1]B, [T v2]B, . . . , [T vn]B] = [[1v1]B, [2v2]B, . . . , [nvn]B] = [1[v1]B, 2[v2]B, . . . , n[vn]B] = [1e1, 2e2, . . . , nen] = diag {1, 2, . . . , n} where, in this computation we used successively: the definition of [T]B (line 1), the fact that T(vj ) = jvj (line 2), linearity of coordinates (line 3), and finally the fact that [vj ]B = ej (line 4). (b) (a): Now suppose B = {v1, v2, . . . vn} is a basis for R n , and that [T]B = diag {1, 2, . . . , n}. We must show that each vj is an eigenvector of T. In fact, we’ll see that T vj = jvj for 1 j n. Fix such a j, and note that the j-th column of [T]B is just jej , i.e., (2.1) [T]Bej = jej EIGENVALUES/EIGENVECTORS 7 Now the right-hand side of (2.1) is j [vj ]B which, by linearity of coordinates, is [jvj ]B. Similarly, the left-hand side of (2.1) is [T]B[vj ]B which, by (1.6), is [T vj ]B. Thus [T vj ]B = [jvj ]B, i.e., the both T vj and jvj have the same B-coordinate vectors. Thus T vj = jvj , as desired. 2.7. Definition. Call an n × n matrix diagonalizable if R n has a basis consisting of eigenvectors of this matrix. Warning: Not all square matrices are diagonalizable! For example it’s geometrically clear that shear transfomations of R 2 have (up to multiplication by a non-zero constant) just a single eigenvector, while rotations of R 2 through any angle not an integer multiple of have no eigenvectors at all! We’ll see how to prove this kind of result“analytically” (i.e. without reference to pictures) in the next section. 3. Efficient eigenvalue computation for 2 × 2 matrices We’ll base our computations for 2 × 2 matrices on the notion of determinant. 3.1. Definition. The determinant of the matrix A = a b c d is: det(A) = ad bc The key to all the work of this section is: 3.2. Theorem. For A a 2 × 2 matrix, the system of linear equations Ax = 0 has nontrivial solution if and only if det(A) = 0. Proof. In scalar form the system of equations is: ax1 + bx2 = 0 cx1 + dx2 = 0 Suppose first that neither a nor c is zero. Then we can replace the second equation by c times the first minus a times the second, yielding the equivalent system: ax1 + bx2 = 0 (ad bc)x2 = 0 8 EIGENVALUES/EIGENVECTORS which has a nontrivial solution if and only if ad bc = 0. Suppose a = 0. Then det(A) = bc, which is zero precisely when either b or c is zero. If b = 0 then we are back to one equation in two unknowns, so there are nontrivial (infinitely many) solutions. If b 6= 0 then any solution must have x2 = 0, whereupon the second equation implies cx1 = 0. If c = 0 (i.e. det(A) = 0) then there are infinitely many solutions: x2 = 0 and x1 arbitrary. If, however c 6= 0 (i.e., det(A) 6= 0) then x1 is also required to be zero, hence the only solution is the trivial one. 3.3. Corollary. The eigenvalues of a 2 × 2 matrix A are the real numbers that satisfy its
  • 5. “characteristic equation” det(a I) = 0. Proof. This follows immediately from the above Theorem and the definition of eigenvalue. More precisely, to say a real number is an eigenvalue of A means that there is a non-zero vector x R 2 such that: Ax = x i.e. Ax x = 0 i.e. (A I)x = 0 Thus is an eigenvector of A if and only if the homogeneous system (A I)x = 0 has a nontrivial solution, which, by Theorem 3.2, happens if and only if det(A I) = 0. 3.4. Example. Show that the matrix A = 1 2 4 3 is diagonalizable. Proof. Of course we know this already from §2.2(b) where the eigenvectors and eigenvalues of A were handed to us. But suppose they weren’t. Then we write the characteristic equation: 0 = det(A I) = det( 1 2 4 3 ) = (1 )(3 ) 8 = 2 4 5 = ( 5)( + 1) EIGENVALUES/EIGENVECTORS 9 so that = 5 and = 1 are the solutions, and so these are the eigenvalues of A. Now for = 5, A I = A 5I = 4 2 4 2 and the system (A 5I)x = 0 is easily seen to have solutions x = any scalar multiple of 1 2 . In particular, v1 = 1 2 is the unique (up to multiplication by a non-zero scalar) eigenvector for the eigenvalue 5. I leave it to you to show by the same method that v2 = 1 1 is the “unique” eigenvector for the eigenvalue 1. It’s easy to check that B = {v1, v2} is a basis for R 2 , so A is diagonalizable, and the matrix of its linear transformation, with respect to the basis B is diag {5, 1}. 3.5. Example: a non-diagonalizable matrix. Let A = 0 1 0 0 . The characteristic equation for this matrix is 2 = 0, which has only = 0 as its solution. The corresponding eigenvectors are the nontrivial solutions to the equation Ax = 0 (i.e. the non-zero vectors in the kernel of A), and these all the non-zero scalar multiples of the standard unit vector e1. Thus there aren’t enough eigenvectors here to make a basis of R 2 , so A is not diagonalizable. 4. Matrix of [T]B vs. standard matrix of T Suppose T is a linear transformation of R n whows “standard matrix” is A, i.e. T(x) = Ax (x R n ). Suppose B = {v1, v2, . . . vn} is a basis for R n . It’s time to get serious about the relationship between A and [T]B. Recall that in §1.8 we derived equation (1.3): [x]B = V 1x where V = [v1, v2, . . . , vn] is the matrix having the vector vj as its j-th column. Recall also that the invertibility of V derives from the fact that its columns are linearly independent. 10 EIGENVALUES/EIGENVECTORS Thus we may compute: [T]B = [[T(v1)]B, [T(v2)]B, . . . , [T(vn)]B] = [[Av1]B, [Av2]B, . . . , [Avn]B] = V 1Av1, V 1Av2, . . . , V 1Avn = V 1A [v1, v2, . . . , vn] = V 1AV where the first line is the definition of [T]B, the second uses the definition of T in terms of A, the third uses (1.3), and the fourth is the definition of matrix multiplication. We have proved: 4.1. Theorem (Fact 3.4.4, Page 144 of our text). Suppose B = {v1, v2, . . . vn} is a basis for R n and T a linear transformation of R n with standard matrix A. Then [T]B = V 1AV , where V is the matrix whose j-th column is vj (1 j n). 5. Application: Powers of matrices Theorem 4.1 is of great utility in computing large powers of matrices—a calculation that would be difficult to do by
  • 6. hand in many circumstances. The method depends on two crucial facts. 5.1. Theorem. If A, B and V are n × n matrices, and V is invertible, then for each positive integer p: (V BV 1 ) p = V BpV 1 . Proof. We use the method of Mathematical Induction, which asserts that to prove the theorem we need only check that it’s true for p = 1, and then show that: (*) Result true for an integer p 1 Result true for p + 1. The point is that once you’ve established the case p = 1 and proved (*), then (*) gives you for free the case p = 2, and then the case p = 3, and so on forever. EIGENVALUES/EIGENVECTORS 11 For our result, the case p = 1 is obviously true, so suppose we have proved the theorem for some p 1. We must show that it’s true for p + 1; to this end write (V BV 1 ) p+1 = (V BV 1 ) p (V BV 1 ) = (V BpV 1 )(V BV 1 ) = (V Bp )(V 1V )(BV 1 ) = (V Bp )I(BV 1 ) = V Bp+1V 1 where the second line follows from our “induction hypothesis” that the theorem is true for p, the third line follows from the associate law of matrix multiplication, and the others are clear. The next result asserts that for a diagonal matrix you can compute a power by taking the corresponding power of each element of the matrix. 5.2. Theorem. For any positive integer p, diag {1, 2, . . . , n} p = diag { p 1 , p 2 , . . . , p n} Proof. I leave this one mostly as an exercise. First check that diag {1, 2, . . . , n}diag {µ1, µ2, . . . , µn} = diag {1µ1, 2µ2, . . . , nµn} The result about powers then follows quickly from this by induction. 5.3. Example. Compute 1 2 4 3 15 .