Command Palette

Search for a command to run...

Linear Methods of AI

Characteristic Polynomial

Definition and Basic Concepts

To find the eigenvalues of a matrix, we need a very important mathematical tool in linear algebra. Imagine we want to find all values λ\lambda that make the matrix AλIA - \lambda I become singular (not invertible).

Let AKn×nA \in \mathbb{K}^{n \times n}. We can form a special function:

χA(t)=det(AtI)\chi_A(t) = \det(A - t \cdot I)
χA(t)=antn+an1tn1++a1t+a0\chi_A(t) = a_n \cdot t^n + a_{n-1} \cdot t^{n-1} + \cdots + a_1 t + a_0

This function is a polynomial of degree nn in tKt \in \mathbb{K}, which we call the characteristic polynomial of AA.

with coefficients a0,,an1,anKa_0, \ldots, a_{n-1}, a_n \in \mathbb{K}.

In fact, χA(t)\chi_A(t) is indeed a polynomial of degree nn for every matrix AKn×nA \in \mathbb{K}^{n \times n}.

Matrix Trace and Polynomial Coefficients

Let AKn×nA \in \mathbb{K}^{n \times n} be a square matrix. The trace of AA is the sum of the diagonal elements:

traceA=i=1naii\text{trace} A = \sum_{i=1}^n a_{ii}

The matrix trace has a close relationship with the coefficients of the characteristic polynomial.

Relationship of Coefficients with Trace and Determinant

In the characteristic polynomial χA(t)\chi_A(t) of AA, its coefficients have special meaning:

an=(1)n,an1=(1)n1traceA,a0=detAa_n = (-1)^n, \quad a_{n-1} = (-1)^{n-1} \cdot \text{trace} A, \quad a_0 = \det A

This means:

  • The highest coefficient is always (1)n(-1)^n
  • The second highest coefficient is related to the matrix trace
  • The constant term is the determinant of the matrix

Eigenvalues as Polynomial Roots

The most important concept of the characteristic polynomial is its relationship with eigenvalues.

Now, let's look at a very important relationship: λK\lambda \in \mathbb{K} is an eigenvalue of AA if and only if det(AλI)=0\det(A - \lambda \cdot I) = 0.

In other words, eigenvalues are the roots of the characteristic polynomial.

The equation for tKt \in \mathbb{K}:

det(AtI)=0\det(A - t \cdot I) = 0

is what we call the characteristic equation of AA.

Algebraic Multiplicity

Now, what happens if an eigenvalue appears multiple times as a root of the characteristic polynomial? Let AKn×nA \in \mathbb{K}^{n \times n} and λK\lambda \in \mathbb{K}. The multiplicity of the root t=λt = \lambda of the characteristic polynomial χA(t)\chi_A(t) is called the algebraic multiplicity μA(λ)\mu_A(\lambda) of the eigenvalue λ\lambda of AA. We say λ\lambda is an eigenvalue with multiplicity μA(λ)\mu_A(\lambda) of AA.

Multiplicity Bounds

For every eigenvalue λ\lambda, it holds:

0μA(λ)n0 \leq \mu_A(\lambda) \leq n

Relationship between Geometric and Algebraic Multiplicity

One important result in eigenvalue theory is the relationship between geometric and algebraic multiplicity.

For any matrix AKn×nA \in \mathbb{K}^{n \times n} and λK\lambda \in \mathbb{K}, we have an interesting relationship:

0dimEigA(λ)μA(λ)n0 \leq \dim \text{Eig}_A(\lambda) \leq \mu_A(\lambda) \leq n

The geometric multiplicity of every eigenvalue is always less than or equal to its algebraic multiplicity.

Why does this happen? This can be explained using basis transformation and Jordan block form.

Examples of Characteristic Polynomial Calculation

After understanding the basic concepts, let's see how to apply them in concrete examples of characteristic polynomial calculation:

3x3 Matrix Example

Let A=(321104301)K3×3A = \begin{pmatrix} 3 & 2 & -1 \\ 1 & 0 & -4 \\ 3 & 0 & 1 \end{pmatrix} \in \mathbb{K}^{3 \times 3}. The characteristic polynomial of AA is:

χA(t)=det(3t211t4301t)\chi_A(t) = \det \begin{pmatrix} 3-t & 2 & -1 \\ 1 & -t & -4 \\ 3 & 0 & 1-t \end{pmatrix}
=(3t)(t)(1t)2(1(1t)+34)+13t= (3-t) \cdot (-t) \cdot (1-t) - 2 \cdot (1 \cdot (1-t) + 3 \cdot 4) + 1 \cdot 3 \cdot t
=3t+3t2+t2t32+2t243t= -3t + 3t^2 + t^2 - t^3 - 2 + 2t - 24 - 3t
=t3+4t24t26= -t^3 + 4t^2 - 4t - 26

For K=R\mathbb{K} = \mathbb{R}, χA(t)\chi_A(t) only has the root λ11.8003\lambda_1 \approx -1.8003 with algebraic multiplicity μA(λ1)=1\mu_A(\lambda_1) = 1.

For K=C\mathbb{K} = \mathbb{C}, χA(t)\chi_A(t) has roots λ11.8003\lambda_1 \approx -1.8003, λ22.9001+2.4559i\lambda_2 \approx 2.9001 + 2.4559i, and λ32.90012.4559i\lambda_3 \approx 2.9001 - 2.4559i with algebraic multiplicities μA(λ1)=μA(λ2)=μA(λ3)=1\mu_A(\lambda_1) = \mu_A(\lambda_2) = \mu_A(\lambda_3) = 1 respectively.

Simple Example

The characteristic polynomial of matrix A=(1201)A = \begin{pmatrix} 1 & 2 \\ 0 & 1 \end{pmatrix} is:

χA(t)=det(AtI)=det(1t201t)\chi_A(t) = \det(A - t \cdot I) = \det \begin{pmatrix} 1-t & 2 \\ 0 & 1-t \end{pmatrix}
=(1t)(1t)02=(1t)2= (1-t) \cdot (1-t) - 0 \cdot 2 = (1-t)^2
=t22t+1= t^2 - 2 \cdot t + 1

This matrix has the root λ=1\lambda = 1 with algebraic multiplicity μA(1)=2\mu_A(1) = 2. λ=1\lambda = 1 is the only eigenvalue of AA. We have calculated that dimEigA(1)=1\dim \text{Eig}_A(1) = 1.

Geometric Transformation Examples

Now, let's explore something fascinating: how the characteristic polynomial works on common geometric transformations we often encounter in R2R2\mathbb{R}^2 \to \mathbb{R}^2:

Rotation

Rotation with A=(cos(α)sin(α)sin(α)cos(α))A = \begin{pmatrix} \cos(\alpha) & -\sin(\alpha) \\ \sin(\alpha) & \cos(\alpha) \end{pmatrix} has the characteristic polynomial:

χA(t)=(cos(α)t)2+sin(α)2=t22cos(α)t+1\chi_A(t) = (\cos(\alpha) - t)^2 + \sin(\alpha)^2 = t^2 - 2 \cdot \cos(\alpha) \cdot t + 1

This has real roots if and only if cos(α)210\cos(\alpha)^2 - 1 \geq 0, that is cos(α)2=1\cos(\alpha)^2 = 1, so α=0\alpha = 0 or α=π\alpha = \pi.

Reflection

Reflection along axes with A=(1001)A = \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix} has the characteristic polynomial:

χA(t)=(1t)(1t)=t21\chi_A(t) = (1-t) \cdot (-1-t) = t^2 - 1

The eigenvalues are λ1=1\lambda_1 = 1 and λ2=1\lambda_2 = -1 with μA(1)=μA(1)=1\mu_A(1) = \mu_A(-1) = 1.

Scaling

Scaling with A=(s00s)A = \begin{pmatrix} s & 0 \\ 0 & s \end{pmatrix} has the characteristic polynomial:

χA(t)=(st)2=t22st+s2\chi_A(t) = (s-t)^2 = t^2 - 2 \cdot s \cdot t + s^2

The eigenvalue is λ=s\lambda = s with μA(s)=2\mu_A(s) = 2.

Shearing

Shearing with A=(1s01)A = \begin{pmatrix} 1 & s \\ 0 & 1 \end{pmatrix} with s0s \neq 0 has the characteristic polynomial:

χA(t)=(1t)2=t22t+1\chi_A(t) = (1-t)^2 = t^2 - 2 \cdot t + 1

The eigenvalue is λ=1\lambda = 1 with μA(1)=2\mu_A(1) = 2.

Properties of Similar Matrices

Similar matrices have a fascinating property: they have the same characteristic polynomial, and therefore have the same eigenvalues, the same trace, and the same determinant.

Let's see why this is true. Let SKn×nS \in \mathbb{K}^{n \times n} be invertible and B=S1ASB = S^{-1} \cdot A \cdot S. Then:

χB(t)=det(BtI)=det(S1AStS1IS)\chi_B(t) = \det(B - t \cdot I) = \det(S^{-1} \cdot A \cdot S - t \cdot S^{-1} \cdot I \cdot S)
=det(S1(AtI)S)= \det(S^{-1} \cdot (A - t \cdot I) \cdot S)
=detS1det(AtI)detS=χA(t)= \det S^{-1} \cdot \det(A - t \cdot I) \cdot \det S = \chi_A(t)

Eigenvector Properties of Similar Matrices

Now, what about the eigenvectors of similar matrices? Let A,BKn×nA, B \in \mathbb{K}^{n \times n} be similar matrices with B=S1ASB = S^{-1} \cdot A \cdot S and invertible matrix SKn×nS \in \mathbb{K}^{n \times n}. If λK\lambda \in \mathbb{K} is an eigenvalue of both AA and BB, and vKnv \in \mathbb{K}^n is an eigenvector of AA for eigenvalue λ\lambda, then w=S1vw = S^{-1} \cdot v is an eigenvector of BB for eigenvalue λ\lambda.

Let's see why this is true. Let Av=λvA \cdot v = \lambda \cdot v and w=S1vw = S^{-1} \cdot v. Then:

Bw=S1ASS1v=S1AvB \cdot w = S^{-1} \cdot A \cdot S \cdot S^{-1} \cdot v = S^{-1} \cdot A \cdot v
=S1λv=λS1v=λw= S^{-1} \cdot \lambda \cdot v = \lambda \cdot S^{-1} \cdot v = \lambda \cdot w

This shows that similarity transformation not only preserves eigenvalues, but also provides a systematic way to transform eigenvectors.