Laurent Hoeltgen

Krylov Subspaces

2021-07-04, update: 2021-08-17 ·  4 min read  ·  Mathematics

This is the first post in a (planned) series on Krylov subspaces, projection processes, and related algorithms. We already discussed projection processes when talking about the Bregman algorithm and we will see that the Krylov (sub-)spaces will be generated by a set of vectors that are not necessarily orthogonal. Getting an orthogonal basis for these spaces can for example be done by applying Gram-Schmidt. Forthcoming posts will discuss further interesting applications such as the Arnoldi iteration or the Lanczos method.

Krylov Subspaces

Krylov subspaces are probably best known for their use in solving linear systems of equations. However, this was not exactly what Aleksey Nikolaevich Krylov had in mind when he published his paper [1] in 1931. He was interested in oscillations of mechanical systems (see [2] for an overview on his scientific achievements). Actually he worked on linear differential equations at that time. Linear systems of equations were basically just one discretisation step away. In his publication [1] Aleksey Krylov proposed a method to reformulate the characteristic equation of a matrix such that computations get easier. The characteristic equation of a matrix AA is given by det(λIA)=0\det(\lambda I - A) = 0 and its solutions tell you a lot about the linear mapping AA.

Krylov's approach is based on the following observation: Assume we have a square matrix ARn,nA\in\mathbb{R}^{n,n} and a non-zero vector vRv\in\mathbb{R}. If we iteratively multiply AA with vv, then we obtain a sequence vv, AvAv, A2vA^2v, .... At some point (at latest after nn steps) this sequence will consist of linearly dependent vectors and adding an additional iterate will not enlarge the spanned space any further. Let's call dd the smallest integer for which we obtain such a set of linearly dependent vectors. Then it follows that there exist coefficients αj\alpha_j such that

Adv=j=0d1αjAjv A^d v = \sum_{j=0}^{d-1} \alpha_j A^j v

We can rewrite this equation as p(A)v=0p(A)v = 0, where p(A)p(A) is given by

Adj=0d1αjAj A^d - \sum_{j=0}^{d-1} \alpha_j A^j

The polynomial pp that we have obtained in this way is called the minimal polynomial of vv with respect to AA.

For the sake of simplicity let us now assume that we have a vector vRnv\in\mathbb{R}^n for which d=nd=n (so dd is as large as possible). We consider the following homogeneous linear system

(vAn1vAnv1λn1λnp(λ))(α0αn11)=(000) \begin{pmatrix} v && \cdots && A^{n-1}v && A^nv \\ 1 && \cdots && \lambda^{n-1} && \lambda^n - p(\lambda) \end{pmatrix} \begin{pmatrix} -\alpha_0 \\ \vdots \\ -\alpha_{n-1} \\ 1 \end{pmatrix} = \begin{pmatrix} 0 \\ \vdots \\ 0 \\ 0 \end{pmatrix}

The first nn equations correspond to p(A)v=0p(A)v = 0. The last equation is just p(λ)p(λ)=0p(\lambda) - p(\lambda) = 0. Since this system has a non-trivial solution, the system matrix must be singular and hence its determinant must be 00. Using the linearity of the determinant in the last row one can derive

p(λ)=det(vAn1vAnv1λn1λn)det(vAn1v) p(\lambda) = \frac{\det\begin{pmatrix} v && \cdots && A^{n-1}v && A^nv \\ 1 && \cdots && \lambda^{n-1} && \lambda^n \end{pmatrix}}{\det\begin{pmatrix} v && \cdots && A^{n-1}v \end{pmatrix}}

Now we just have to observe that p(λ)p(\lambda) is also equal to the characteristic polynomial of AA and we have found our sought expression! The benefit of the previous expression is that λ\lambda only appears in a single line of the determinant in the numerator. For the usual formula, λ\lambda appears in every line.

Of course, certain conditions on AA and vv must apply for the previous formula to apply (we already assumed that d=nd=n above). One of Krylovs observations is (roughly speaking) that for each matrix AA and each vector, the nested sequence of spaces

span{v,Av,,An1v},n=1,2, \mathrm{span}\left\lbrace v, Av, \ldots, A^{n-1}v\right\rbrace,\quad n=1, 2, \ldots

will eventually stop to grow and become invariant under AA. These spaces in the previous equation are nowadays called Krylov spaces. In the forthcoming we will denote them by

Vn(A,v):=span{v,Av,A2v,,An1v} V_n(A, v) := \mathrm{span}\lbrace v, Av, A^2v, \ldots, A^{n-1}v \rbrace

Relationship to Power Iterations

Looking at the definition of Vn(A,v)V_n(A, v) we immediately notice the similarity to the classic power method. The power method starts with a vector w0=vw_0=v and iteratively computes

wk+12=Awk,wk+1=wk+12wk+12 w_{k+\frac{1}{2}} = A w_k,\quad w_{k+1} = \frac{w_{k+\frac{1}{2}}}{\left\lVert w_{k+\frac{1}{2}}\right\rVert}

If certain requirements are met, then the sequence (wk)k(w_k)_k converges towards the eigenvector corresponding to the largest eigenvalue in magnitude. Hence the iterates from the power method form a basis of the Krylov subspace.

Examples

If A=IA=I, i.e. AA is the identity matrix, then, for any vector vv we have

V0(A,v)=V0(I,v)=span{v}=V1(I,v)==Vn(I,v) V_0(A,v) = V_0(I,v) = \mathrm{span}\lbrace v \rbrace = V_1(I,v) = \ldots = V_n(I,v)

for any n>0n>0. In this case all our Krylov subspaces are identical and have dimension 1.

In general however, we have Vk(A,v)Vk+1(A,v)V_k(A,v) \subseteq V_{k+1}(A,v) because you add AkvA^kv to the span. If we consider

A=(842421210),v=(111) A = \begin{pmatrix} 8 & 4 & 2 \\ 4 & 2 & 1 \\ 2 & 1 & 0 \end{pmatrix},\quad v = \begin{pmatrix} 1 \\ 1 \\ 1 \end{pmatrix}

then it's easy to check that V0(A,v)V1(A,v)V_0(A,v) \subset V_1(A,v) and V1(A,v)V2(A,v)=R3V_1(A,v) \subset V_2(A,v) = \mathbb{R}^3.

Estimating the maximal dimension of a Krylov subspace, without explicitly computing all spaces Vk(A,v)V_k(A,v) would be out of scope for this article.

References

[1] A. N. Krylov, О численном решении уравнения, которым в технических вопросах определяются частоты малых колебаний материальных систем (De la résolution numérique de l'équation servant à déterminer dans des questions de mécanique appliquée les fréquences de petites oscillations des systèmes matériels), Izvestiia Akademii nauk SSSR 7(4): 491–539, 1931. In Russian

[2] J. Liesen, Zdenĕk Strakoš, Krylov Subspace Methods: Principles and Analysis; Oxford University Press, 2013

How to cite this page

Hoeltgen, Laurent: Krylov Subspaces, 2021-08-17.

BibLaTeX code:
@online{krylov-spaces,
  author   = {Hoeltgen, Laurent},
  title    = {Krylov Subspaces},
  date     = {2021-08-17},
  language = english
  url      = {https://www.laurenthoeltgen.name/content/blog/
              krylov-spaces}
}
Download BibLaTeX file

CC BY-SA 4.0 Laurent Hoeltgen. Last modified: September 19, 2025.
Website built with Franklin.jl and the Julia programming language.
Privacy Policy · Terms