Eigenvalues and Eigenvectors
Some special vectors survive a matrix transformation with their direction intact -- they only get scaled. These eigenvectors and their eigenvalues reveal the deepest structure of linear transformations, powering everything from Google search to PCA.
The Big Question
Apply a matrix to a vector and almost anything can happen. Most vectors get rotated and stretched, landing in a brand new direction. But scattered among them are a few special vectors that refuse to leave their span. They come out of the transformation pointing exactly the way they went in – only longer, shorter, or flipped.
These survivors are eigenvectors. The factor by which they get scaled is the eigenvalue.
Hunting for these “invariant directions” is one of the most powerful moves in all of mathematics. Eigenvalues drive Google’s PageRank, predict population growth, govern the stability of bridges and aircraft, and turn matrix problems that would take centuries on a supercomputer into a handful of arithmetic.
What you will learn
- The definition of eigenvectors and eigenvalues, and a clean geometric picture of both
- How to find them with the characteristic equation
- Diagonalization: rewriting a matrix as scaling along its own natural axes
- Why complex eigenvalues show up exactly when a matrix rotates
- Power iteration, PCA, PageRank, Leslie populations, Fibonacci – one idea, five flavours
- The spectral theorem for symmetric matrices
Prerequisites
Chapter 3 (matrices as transformations), Chapter 4 (determinants), and Chapter 5 (null space and column space).
What an Eigenvector Really Is
The formal statement
$$A\vec{v} = \lambda\vec{v},$$then $\vec{v}$ is an eigenvector of $A$ and $\lambda$ is the corresponding eigenvalue.
The word “eigen” is German for “own” or “intrinsic.” Eigenvectors are the matrix’s own private set of directions.
Why non-zero matters
The zero vector trivially satisfies $A\vec{0} = \lambda\vec{0}$ for any $\lambda$ and any $A$. If we let it count, every scalar would be an eigenvalue and the concept would carry no information. So we insist that eigenvectors are non-zero – eigenvalues, however, are allowed to be zero (they signal that $A$ collapses that direction onto the origin).
A picture is worth a thousand symbols
The definition becomes obvious once you see it. Pick two vectors in the plane and apply the same matrix $A$ to both. One is on an eigen-line; the other is not.

The green vector $\vec{v}$ comes out three times longer but pointing the same way – its line through the origin is preserved. The amber vector $\vec{u}$ is rotated off its original line. Eigenvectors are precisely the vectors for which the dashed line in the left panel and the dashed line in the right panel coincide.
Two intuition pumps
Kneading dough. When you knead, almost every point inside the dough moves to a totally new location, mixed and rotated. Yet along the rolling pin’s axis, points only get compressed – their direction is preserved. Those are the eigen-directions of that particular squashing.
A spinning ride. On a merry-go-round, every direction in your body keeps changing – except the rotation axis, which always points up. That axis is an eigenvector with eigenvalue 1.
What the Eigenvalue Tells You
The eigenvalue $\lambda$ is purely a scaling factor along the eigen-line. Three cases capture the whole story:

The complete dictionary:
| Eigenvalue | What happens to the eigenvector |
|---|---|
| $\lambda > 1$ | Stretched (longer, same direction) |
| $\lambda = 1$ | Unchanged (a true fixed direction) |
| $0 < \lambda < 1$ | Compressed (shorter, same direction) |
| $\lambda = 0$ | Crushed onto the origin – $A$ is singular along this line |
| $-1 < \lambda < 0$ | Flipped and shrunk |
| $\lambda < -1$ | Flipped and stretched |
| $\lambda \in \mathbb{C}\setminus\mathbb{R}$ | Rotation – no real direction is fixed |
Worked example: the mirror
Reflection across the y-axis is given by $A = \bigl(\begin{smallmatrix}-1&0\\0&1\end{smallmatrix}\bigr)$.
- $\lambda_1 = -1$, eigenvector $(1,0)$: horizontal vectors get flipped left-right.
- $\lambda_2 = 1$, eigenvector $(0,1)$: vertical vectors do not move at all.
That matches your everyday experience of a vertical mirror: left and right swap, up and down stay put.
Worked example: the shear
The shear $A = \bigl(\begin{smallmatrix}1&1\\0&1\end{smallmatrix}\bigr)$ tilts every vertical line into a slanted line. Solving the characteristic equation gives $\lambda = 1$ as a repeated eigenvalue, with the only independent eigenvector $(1,0)$.
In other words, a shear has exactly one direction that survives – the horizontal axis – and no second one. We will see that this kind of matrix, called defective, cannot be diagonalised.
Finding Eigenvalues: the Characteristic Equation
$$(A - \lambda I)\vec{v} = \vec{0}.$$$$\det(A - \lambda I) = 0.$$This is the characteristic equation. It is a polynomial of degree $n$ in $\lambda$, so by the fundamental theorem of algebra it has exactly $n$ roots in $\mathbb{C}$ (counted with multiplicity). Those roots are the eigenvalues.
Full worked example
Find the eigenvalues and eigenvectors of $A = \bigl(\begin{smallmatrix}4&2\\1&3\end{smallmatrix}\bigr)$.
$$\det\begin{pmatrix}4-\lambda&2\\1&3-\lambda\end{pmatrix} = (4-\lambda)(3-\lambda) - 2 = \lambda^2 - 7\lambda + 10 = 0.$$Step 2 – solve. $(\lambda - 5)(\lambda - 2) = 0$, so $\lambda_1 = 5$ and $\lambda_2 = 2$.
Step 3 – eigenvectors.
$$(A - 5I)\vec{v} = \begin{pmatrix}-1&2\\1&-2\end{pmatrix}\vec{v} = \vec{0}\;\Longrightarrow\; v_1 = 2v_2,\;\; \vec{v}_1 = (2, 1).$$$$(A - 2I)\vec{v} = \begin{pmatrix}2&2\\1&1\end{pmatrix}\vec{v} = \vec{0}\;\Longrightarrow\; v_1 = -v_2,\;\; \vec{v}_2 = (-1, 1).$$Verification. $A(2,1)^T = (10,5)^T = 5\,(2,1)^T$. ✓
Two free sanity checks
$$\operatorname{tr}(A) = \lambda_1 + \cdots + \lambda_n,\qquad \det(A) = \lambda_1\cdots\lambda_n.$$For our example: $\operatorname{tr}(A) = 7 = 5+2$ and $\det(A) = 10 = 5\cdot 2$. Always check this – it catches arithmetic mistakes in seconds.
Visualising eigenvectors of a $2\times 2$ matrix
Plotting the original grid alongside the deformed grid makes the eigenvectors leap out: they are the lines along which the grid hinges.

The grid in the right panel is sheared and stretched, but the green and purple lines through the eigenvectors stay exactly where they were. Every other line tilts.
Diagonalization: Making a Matrix Trivial
The core idea
$$A = P D P^{-1}.$$The transformation $A$ has been factored into three pieces: change to the eigenbasis, scale each eigen-axis independently, change back.
Picture: three steps that compose to $A$

In the middle panel the transformation is just $D$ – you stretch the x-axis by 3 and leave the y-axis alone. In the eigenbasis, every diagonalisable matrix is just a list of independent stretches.
Why this is wonderful
Cheap matrix powers. Because $A^k = P D^k P^{-1}$ and $D^k = \operatorname{diag}(\lambda_1^k, \ldots, \lambda_n^k)$, raising $A$ to the 100th power costs almost nothing once the eigendecomposition is known – you raise a few scalars, not matrices.
Long-term behaviour. As $k \to \infty$, the dominant eigenvalue (the largest in absolute value) controls everything. If $|\lambda_{\max}| > 1$ the iterated system explodes; if $|\lambda_{\max}| < 1$ it dies; if $|\lambda_{\max}| = 1$ it stays bounded.
Function calculus on matrices. Define $f(A) := P f(D) P^{-1}$ for any reasonable function – this is how matrix exponentials in differential equations and continuous-time Markov chains work.
When can you diagonalise?
- $n$ distinct eigenvalues guarantee diagonalisability (eigenvectors of distinct eigenvalues are automatically independent – a fact we will prove later).
- Real symmetric matrices (and more generally, normal matrices) are always diagonalisable, even with repeated eigenvalues, and even with an orthonormal basis of eigenvectors.
- A matrix is defective if it has fewer independent eigenvectors than its size. The shear $\bigl(\begin{smallmatrix}1&1\\0&1\end{smallmatrix}\bigr)$ is the canonical example: $\lambda = 1$ has algebraic multiplicity 2 but geometric multiplicity 1.
Complex Eigenvalues: the Mathematics of Rotation
Rotations have no real eigenvectors
$$A = \begin{pmatrix}0&-1\\1&0\end{pmatrix},$$with characteristic equation $\lambda^2 + 1 = 0$, giving $\lambda = \pm i$. There is no real eigenvector, and that is geometrically obvious: a 90-degree rotation moves every direction in the real plane.

The left panel shows that no matter which direction you try, $R$ kicks it off its line. The right panel iterates $R$ on a single starting vector – the trajectory traces a circle, with constant length, exactly because $|\lambda| = |e^{i\theta}| = 1$.
General rotation
$$R_\theta = \begin{pmatrix}\cos\theta&-\sin\theta\\\sin\theta&\cos\theta\end{pmatrix},\qquad \lambda = \cos\theta \pm i\sin\theta = e^{\pm i\theta}.$$That is Euler’s formula falling out of a determinant. The modulus $|\lambda| = 1$ encodes the fact that rotation preserves length.
Complex eigenvalues come in conjugate pairs
Real matrices have real characteristic polynomials, so complex roots arrive as conjugates: if $\lambda = a + bi$ is an eigenvalue, then so is $\bar\lambda = a - bi$.
What complex eigenvalues mean for a system
Write $\lambda = re^{i\theta}$. Each iteration of $A$ scales by $r$ and rotates by $\theta$. So the long-term behaviour of $A^k\vec{x}_0$ is
- $r < 1$: spiral inward (damped oscillation),
- $r = 1$: orbit on a circle/ellipse (pure oscillation),
- $r > 1$: spiral outward (growing oscillation).
This is exactly why an undamped pendulum swings forever, why a damped one settles, and why a slightly mistuned feedback loop in control theory blows up.
Power Iteration: Finding the Dominant Eigenvector
$$\vec{v}_{k+1} = \frac{A\vec{v}_k}{\|A\vec{v}_k\|}.$$Start with anything (almost any starting vector works), apply $A$, normalise, repeat. The iterates converge to the dominant eigenvector.

After $k$ iterations the term with the largest $|\lambda_i|$ dominates the rest by the factor $(|\lambda_1|/|\lambda_2|)^k$. Normalising kills the absolute scale, and what remains is the dominant eigenvector. Convergence is geometric in the spectral gap $|\lambda_2|/|\lambda_1|$ – bigger gap, faster convergence.
| |
For all eigenvalues, the production-quality algorithm is QR iteration, which NumPy, MATLAB, and LAPACK use under the hood:
| |
Application 1: Population Growth (Leslie Matrix)
Split a species into age classes; let $\vec{p}_t$ count the population in each class at time $t$. The Leslie matrix $L$ has fertilities along the top row and survival probabilities on the sub-diagonal, and the model is just $\vec{p}_{t+1} = L\vec{p}_t$.
$$L = \begin{pmatrix}0&2&0.5\\0.6&0&0\\0&0.8&0\end{pmatrix}.$$ | |
The dominant eigenvalue decides everything: $|\lambda_1| > 1$ means the population grows; $|\lambda_1| < 1$ means it goes extinct; $|\lambda_1| = 1$ means it stabilises. The corresponding eigenvector is the stable age distribution – whatever ratios you start with, you eventually converge to this one.
Application 2: Google PageRank
Larry Page and Sergey Brin asked: how do you measure a page’s importance? Their answer: a page is important if important pages link to it. That definition is recursive, and recursion of this shape is solved by an eigenvector.
$$\vec{r} = H\vec{r},$$which is exactly $H\vec{r} = 1\cdot\vec{r}$ – the eigenvector for eigenvalue 1.
$$G = dH + \frac{1-d}{n}\,\mathbf{1}\mathbf{1}^T.$$Power iteration on $G$ then converges to PageRank. Operationally PageRank is power iteration on a sparse matrix with billions of rows.
| |
Application 3: Fibonacci and the Golden Ratio
$$\begin{pmatrix}F_{n+1}\\F_n\end{pmatrix} = \begin{pmatrix}1&1\\1&0\end{pmatrix}^n \begin{pmatrix}1\\0\end{pmatrix}.$$$$\phi = \tfrac{1+\sqrt 5}{2}\approx 1.618 \quad\text{(the golden ratio)},\qquad \hat\phi = \tfrac{1-\sqrt 5}{2}\approx -0.618.$$$$F_n = \frac{\phi^n - \hat\phi^n}{\sqrt 5}.$$Since $|\hat\phi| < 1$, the second term decays and $F_{n+1}/F_n \to \phi$. The golden ratio is not a coincidence – it is the dominant eigenvalue of the Fibonacci matrix.
Application 4: PCA Preview
Take a cloud of data points in $\mathbb{R}^n$, centre it, and compute the covariance matrix $C = \tfrac{1}{n-1}XX^T$. The eigenvectors of $C$ are the principal components – the orthogonal directions of greatest variance. The eigenvalues are the variances along those directions.

PC$_1$ points along the longest axis of the cloud; PC$_2$ is perpendicular and shorter. Project onto the first few PCs and you get the lowest-dimensional representation that preserves the most information. That single idea powers face recognition, gene-expression clustering, recommendation systems, and – as we will see in Chapter 9 – the singular value decomposition.
Symmetric Matrices: the Spectral Theorem
Real symmetric matrices ($A = A^T$) are the friendliest matrices in linear algebra:
- All eigenvalues are real (no complex roots).
- Eigenvectors of distinct eigenvalues are orthogonal.
- There is always an orthonormal basis of eigenvectors. Equivalently, $A = Q\Lambda Q^T$ with $Q$ orthogonal ($Q^{-1} = Q^T$).
This is the spectral theorem, and it is the reason that PCA, image compression, and most of optimisation work as cleanly as they do.
Spectral decomposition
$$A = \lambda_1\,\vec{q}_1\vec{q}_1^T + \lambda_2\,\vec{q}_2\vec{q}_2^T + \cdots + \lambda_n\,\vec{q}_n\vec{q}_n^T.$$Truncating the sum after the $k$ largest $|\lambda_i|$ gives the best rank-$k$ approximation to $A$ in the Frobenius norm – the engine of low-rank denoising and compression.
Positive definiteness
For a symmetric $A$, the signs of the eigenvalues classify its quadratic form:
| Condition | Quadratic form $\vec{x}^T A\vec{x}$ |
|---|---|
| All $\lambda_i > 0$ | Positive definite – bowl with a unique minimum |
| All $\lambda_i \ge 0$ | Positive semi-definite |
| All $\lambda_i < 0$ | Negative definite – dome with a unique maximum |
| Mixed signs | Indefinite – saddle |
In optimisation, the Hessian of an objective at a critical point being positive definite is the second-order condition for a strict local minimum.
Properties Worth Memorising
| Object | Eigenvalues | Eigenvectors |
|---|---|---|
| $A$ | $\lambda_i$ | $\vec{v}_i$ |
| $A^k$ | $\lambda_i^k$ | $\vec{v}_i$ (same) |
| $A^{-1}$ (if invertible) | $1/\lambda_i$ | $\vec{v}_i$ (same) |
| $A + cI$ | $\lambda_i + c$ | $\vec{v}_i$ (same) |
| $cA$ | $c\lambda_i$ | $\vec{v}_i$ (same) |
| $A^T$ (real) | $\lambda_i$ | generally different |
Similar matrices
If $B = P^{-1}AP$ then $A$ and $B$ are similar, and they have the same characteristic polynomial – hence the same eigenvalues with the same multiplicities. Diagonalisation is just the lucky case where the similarity makes $B$ diagonal.
Algebraic vs. geometric multiplicity
- Algebraic multiplicity of $\lambda$: how many times it appears as a root of the characteristic polynomial.
- Geometric multiplicity of $\lambda$: $\dim \ker(A - \lambda I)$, the number of independent eigenvectors for that $\lambda$.
The geometric multiplicity is always $\le$ the algebraic multiplicity. A matrix is diagonalisable iff equality holds for every eigenvalue.
Chapter Summary
Eigenvalues and eigenvectors uncover the skeleton of a linear transformation:
| Concept | What it tells you |
|---|---|
| Eigenvector | A direction whose span survives the transformation |
| Eigenvalue | The scaling factor along that direction |
| Diagonalisation | Rewriting the transformation as independent stretches in the eigenbasis |
| Complex eigenvalues | The matrix rotates – no real direction is fixed |
| Dominant eigenvalue | Controls long-term behaviour of $A^k\vec{x}_0$ |
| Spectral theorem | Symmetric $\Rightarrow$ real eigenvalues + orthonormal eigenvectors |
Core intuition. Find the directions whose direction does not change, and a complicated transformation reduces to a list of simple stretches.
From Google search to population biology, from quantum mechanics to deep learning – eigenvalues are everywhere. Master this idea and you hold one of the master keys to applied mathematics.
What Comes Next
- Chapter 7: Orthogonality and projections – the geometry that makes PCA and least squares possible.
- Chapter 8: Symmetric matrices and quadratic forms – a deep dive into the spectral theorem.
- Chapter 9: Singular value decomposition – every matrix is rotation × scaling × rotation, even non-square ones.
- Chapters 10-18: Norms, sparse methods, matrix calculus, and applications across ML, vision, and beyond.
Series Navigation
- Previous: Chapter 5 – Linear Systems and Column Space
- Next: Chapter 7 – Orthogonality (coming soon)
- Series: Essence of Linear Algebra (6 of 18)