Eigenvalues and Eigenvectors
Basic definition and theorems
Definition. Let V be a vector space and A:V→V a linear
transformation. If v∈V and λ∈F satisfy
Av=λv,v=0
then v is called an eigenvector of A and λ is called an eigenvalue of A.
Theorem. If A:V→V is a linear transformation then
- v∈V is an eigenvector of A with eigenvalue λ ⟺ v=0 and v∈N(A−λI)
- λ∈F is an eigenvalue of A if and only if N(A−λI)={0}
- λ∈F is an eigenvalue of A if and only if det(A−λI)=0
This theorem gives a method of finding all eigenvalues&vectors, namely:
- Compute the characteristic polynomial det(A−λI)
- Solve det(A−λI)=0 for λ. Let λ1, …,
λk be all the solutions you find
- For each i=1,…,k find all the vectors in N(A−λiI), i.e.
solve Av−λiv=0 for v.
Independence of Eigenvectors Theorem. If v1,…,vm∈V are
eigenvectors of A:V→V with eigenvalues λ1,…,λm, and
if the eigenvalues are distinct (i.e. λi=λj for all i=j) then {v1,…,vm} is linearly independent.
Proof. Suppose {v1,…,vm} is dependent, i.e.~there are numbers c1,…,cm∈F for which c1v1+⋯+cmvm=0, while not all ci vanish. After renumbering the terms we may assume that c1=0.
Consider the linear transformation
B=(A−λ2I)(A−λ3I)⋯(A−λmI)
Then Bvi=0 for all i=2,3,…,m, while
Bv1=(λ1−λ2)(λ1−λ3)⋯(λ1−λm)v1=0.
Applying B to both sides in the equation c1v1+⋯+cmvm=0, we get
0=B(c1v1+c2v2+⋯+cmvm)=c1Bv1+c2Bv2+⋯+cmBvm=c1(λ1−λ2)(λ1−λ3)⋯(λ1−λm)v1.
Since v1=0, and since λ1−λi=0 for all i≥2, it follows that c1=0.
The use of eigenvalues and vectors
If a vector v∈V is known as a linear combination of eigenvectors of A then it is easy to compute Av, A2v, etc. If
v=a1v1+⋯+akvk
where Avi=λivi, then
Av=λ1a1v1+⋯+λkakvkA2v=λ12a1v1+⋯+λk2akvk⋮Aℓv=λ1ℓa1v1+⋯+λkℓakvk
for any ℓ∈N. If none of the eigenvalues λi vanishes then one solution of Aw=v is given by
w=λ1a1v1+⋯+λkakvk.
For this to be useful we need to find as many eigenvectors as we can.
Finding eigenvalues/vectors for A:Fn→Fn
The characteristic polynomial of A:Fn→Fn is, by definition,
det(A−λI)=∣∣∣∣∣∣∣∣∣a11−λa21a11a12a22−λa12⋯⋯⋱⋯a1na2nann−λ∣∣∣∣∣∣∣∣∣
After expanding, this turns out to be a polynomial in λ of degree n
det(A−λI)=(−λ)n+c1(−λ)n−1+c2(−λ)n−2+⋯+cn−1(−λ)+cn.
Here
c1=a11+a22+⋯+ann
is called the trace of the matrix A, and
cn=detA
is its determinant.
Examples
|
|
|
|
|
Some 2×2 matrix |
A=(22−27) |
|
The Fibonacci matrix |
F=(1110) |
Rotation by 90∘ |
R=(01−10) |
|
Reflection |
S=(0110) |
Projection |
P=(1/21/21/21/2) |
|
3D rotation |
R=(001100010) |
Shear |
Z=(1011) |
|
|
|
|
|
|
|
|
Diagonalization
Since the characteristic polynomial det(A−λI) is a polynomial of
degree n, it has at most n zeroes. If it has n distinct zeroes,
λ1, … , λn then we choose an eigenvector vi for each
eigenvalue λi. The vectors v1,…,vn are linearly independent
in Fn and therefore they form a basis for Fn.
Diagonalization Theorem (version 1). If a linear transformation A:V→V has a basis v1,…,vn∈V of eigenvectors, with corresponding eigenvalues λ1, … , λn, then the matrix of A with respect to this basis is
[A]v1,…,vn=⎝⎜⎜⎜⎜⎜⎛λ10000λ20000λ30………⋱…000λn⎠⎟⎟⎟⎟⎟⎞
This follows directly from the fact that Avi=λivi for i=1,2,…,n.
Diagonalization Theorem (version 2). If an n×n matrix A has a basis v1,…,vn∈Fn of eigenvectors, with corresponding eigenvalues λ1, … , λn, then we have
S−1AS=D,
where
S=[v1v2⋯vn]
is the matrix whose columns are the eigenvectors, and D is the diagonal matrix
D=⎝⎜⎜⎜⎜⎜⎛λ10000λ20000λ30………⋱…000λn⎠⎟⎟⎟⎟⎟⎞
Definition. Two linear transformations A:V→V and B:W→W are called similar if there is an invertible linear transformation S:V→W such that
SA=BS, or A=S−1BS, or SAS−1=B.
The three conditions are equivalent.
Solving cubic equations