Conjugate gradient method

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search
A comparison of the convergence of gradient descent with optimal step size (in green) and conjugate vector (in red) for minimizing a quadratic function associated with a given linear system. Conjugate gradient, assuming exact arithmetic, converges in at most n steps, where n is the size of the matrix of the system (here n = 2).

In mathematics, the conjugate gradient method is an algorithm for the numerical solution of particular systems of linear equations, namely those whose matrix is positive-semidefinite. The conjugate gradient method is often implemented as an iterative algorithm, applicable to sparse systems that are too large to be handled by a direct implementation or other direct methods such as the Cholesky decomposition. Large sparse systems often arise when numerically solving partial differential equations or optimization problems.

The conjugate gradient method can also be used to solve unconstrained optimization problems such as energy minimization. It is commonly attributed to Magnus Hestenes and Eduard Stiefel,[1][2] who programmed it on the Z4,[3] and extensively researched it.[4][5]

The biconjugate gradient method provides a generalization to non-symmetric matrices. Various nonlinear conjugate gradient methods seek minima of nonlinear optimization problems.

Description of the problem addressed by conjugate gradients

[edit | edit source]

Suppose we want to solve the system of linear equations

๐€๐ฑ=๐›

for the vector ๐ฑ, where the known nร—n matrix ๐€ is symmetric (i.e., ๐€๐–ณ=๐€), positive-definite (i.e. ๐ฑ๐–ณ๐€๐ฑ>0 for all non-zero vectors ๐ฑ in โ„n), and real, and ๐› is known as well. We denote the unique solution of this system by ๐ฑ*.

Derivation as a direct method

[edit | edit source]

The conjugate gradient method can be derived from several different perspectives, including specialization of the conjugate direction method for optimization, and variation of the Arnoldi/Lanczos iteration for eigenvalue problems. Despite differences in their approaches, these derivations share a common topicโ€”proving the orthogonality of the residuals and conjugacy of the search directions. These two properties are crucial to developing the well-known succinct formulation of the method.

We say that two non-zero vectors ๐ฎ and ๐ฏ are conjugate (with respect to ๐€) if

๐ฎ๐–ณ๐€๐ฏ=0.

Since ๐€ is symmetric and positive-definite, the left-hand side defines an inner product

๐ฎ๐–ณ๐€๐ฏ=โŸจ๐ฎ,๐ฏโŸฉ๐€:=โŸจ๐€๐ฎ,๐ฏโŸฉ=โŸจ๐ฎ,๐€๐–ณ๐ฏโŸฉ=โŸจ๐ฎ,๐€๐ฏโŸฉ.

Two vectors are conjugate if and only if they are orthogonal with respect to this inner product. Being conjugate is a symmetric relation: if ๐ฎ is conjugate to ๐ฏ, then ๐ฏ is conjugate to ๐ฎ. Suppose that

P={๐ฉ1,,๐ฉn}

is a set of n mutually conjugate vectors with respect to ๐€, i.e. ๐ฉi๐–ณ๐€๐ฉj=0 for all iโ‰ j. Then P forms a basis for โ„n, and we may express the solution ๐ฑ* of ๐€๐ฑ=๐› in this basis:

๐ฑ*=โˆ‘i=1nฮฑi๐ฉiโ‡’๐€๐ฑ*=โˆ‘i=1nฮฑi๐€๐ฉi.

Left-multiplying the problem ๐€๐ฑ=๐› with the vector ๐ฉk๐–ณ yields

๐ฉk๐–ณ๐›=๐ฉk๐–ณ๐€๐ฑ*=โˆ‘i=1nฮฑi๐ฉk๐–ณ๐€๐ฉi=โˆ‘i=1nฮฑiโŸจ๐ฉk,๐ฉiโŸฉ๐€=ฮฑkโŸจ๐ฉk,๐ฉkโŸฉ๐€

and so

ฮฑk=โŸจ๐ฉk,๐›โŸฉโŸจ๐ฉk,๐ฉkโŸฉ๐€.

This gives the following method[4] for solving the equation ๐€๐ฑ=๐›: find a sequence of n conjugate directions, and then compute the coefficients ฮฑk.

As an iterative method

[edit | edit source]

If we choose the conjugate vectors ๐ฉk carefully, then we may not need all of them to obtain a good approximation to the solution ๐ฑ*. So, we want to regard the conjugate gradient method as an iterative method. This also allows us to approximately solve systems where n is so large that the direct method would take too much time.

We denote the initial guess for ๐ฑ* by ๐ฑ0 (we can assume without loss of generality that ๐ฑ0=๐ŸŽ, otherwise consider the system ๐€๐ณ=๐›โˆ’๐€๐ฑ0 instead). Starting with ๐ฑ0 we search for the solution and in each iteration we need a metric to tell us whether we are closer to the solution ๐ฑ* (that is unknown to us). This metric comes from the fact that the solution ๐ฑ* is also the unique minimizer of the following quadratic function

f(๐ฑ)=12๐ฑ๐–ณ๐€๐ฑโˆ’๐ฑ๐–ณ๐›,๐ฑโˆˆโ„n.

The existence of a unique minimizer is apparent as its Hessian matrix of second derivatives is symmetric positive-definite

๐‡(f(๐ฑ))=๐€,

and that the minimizer (use Df(๐ฑ)=0) solves the initial problem follows from its first derivative

โˆ‡f(๐ฑ)=๐€๐ฑโˆ’๐›.

This suggests taking the first basis vector ๐ฉ0 to be the negative of the gradient of f at ๐ฑ=๐ฑ0. The gradient of f equals ๐€๐ฑโˆ’๐›. Starting with an initial guess ๐ฑ0, this means we take ๐ฉ0=๐›โˆ’๐€๐ฑ0. The other vectors in the basis will be conjugate to the gradient, hence the name conjugate gradient method. Note that ๐ฉ0 is also the residual provided by this initial step of the algorithm.

Let ๐ซk be the residual at the kth step:

๐ซk=๐›โˆ’๐€๐ฑk.

As observed above, ๐ซk is the negative gradient of f at ๐ฑk, so the gradient descent method would require to move in the direction rk. Here, however, we insist that the directions ๐ฉk must be conjugate to each other. A practical way to enforce this is by requiring that the next search direction be built out of the current residual and all previous search directions. The conjugation constraint is an orthonormal-type constraint and hence the algorithm can be viewed as an example of Gram-Schmidt orthonormalization. This gives the following expression:

๐ฉk=๐ซkโˆ’โˆ‘i<k๐ซk๐–ณ๐€๐ฉi๐ฉi๐–ณ๐€๐ฉi๐ฉi

(see the picture at the top of the article for the effect of the conjugacy constraint on convergence). Following this direction, the next optimal location is given by

๐ฑk+1=๐ฑk+ฮฑk๐ฉk

with

ฮฑk=๐ฉk๐–ณ(๐›โˆ’๐€๐ฑk)๐ฉk๐–ณ๐€๐ฉk=๐ฉk๐–ณ๐ซk๐ฉk๐–ณ๐€๐ฉk,

where the last equality follows from the definition of ๐ซk . The expression for ฮฑk can be derived if one substitutes the expression for xk+1 into f and minimizing it with respect to ฮฑk

f(๐ฑk+1)=f(๐ฑk+ฮฑk๐ฉk)=:g(ฮฑk)g(ฮฑk)=!0โ‡’ฮฑk=๐ฉk๐–ณ(๐›โˆ’๐€๐ฑk)๐ฉk๐–ณ๐€๐ฉk.

The resulting algorithm

[edit | edit source]

The above algorithm gives the most straightforward explanation of the conjugate gradient method. Seemingly, the algorithm as stated requires storage of all previous searching directions and residue vectors, as well as many matrixโ€“vector multiplications, and thus can be computationally expensive. However, a closer analysis of the algorithm shows that ๐ซi is orthogonal to ๐ซj, i.e. ๐ซi๐–ณ๐ซj=0, for iโ‰ j. And ๐ฉi is ๐€-orthogonal to ๐ฉj, i.e. ๐ฉi๐–ณ๐€๐ฉj=0, for iโ‰ j. This can be regarded that as the algorithm progresses, ๐ฉi and ๐ซi span the same Krylov subspace, where ๐ซi form the orthogonal basis with respect to the standard inner product, and ๐ฉi form the orthogonal basis with respect to the inner product induced by ๐€. Therefore, ๐ฑk can be regarded as the projection of ๐ฑ on the Krylov subspace.

That is, if the CG method starts with ๐ฑ0=0, then[6]xk=argminyโˆˆโ„n{(xโˆ’y)โŠคA(xโˆ’y):yโˆˆspan{b,Ab,โ€ฆ,Akโˆ’1b}}The algorithm is detailed below for solving ๐€๐ฑ=๐› where ๐€ is a real, symmetric, positive-definite matrix. The input vector ๐ฑ0 can be an approximate initial solution or ๐ŸŽ. It is a different formulation of the exact procedure described above.

๐ซ0:=๐›โˆ’๐€๐ฑ0if ๐ซ0 is sufficiently small, then return ๐ฑ0 as the result๐ฉ0:=๐ซ0k:=0repeatฮฑk:=๐ซk๐–ณ๐ซk๐ฉk๐–ณ๐€๐ฉk๐ฑk+1:=๐ฑk+ฮฑk๐ฉk๐ซk+1:=๐ซkโˆ’ฮฑk๐€๐ฉkif ๐ซk+1 is sufficiently small, then exit loopฮฒk:=๐ซk+1๐–ณ๐ซk+1๐ซk๐–ณ๐ซk๐ฉk+1:=๐ซk+1+ฮฒk๐ฉkk:=k+1end repeatreturn ๐ฑk+1 as the result

This is the most commonly used algorithm. The same formula for ฮฒk is also used in the Fletcherโ€“Reeves nonlinear conjugate gradient method.

Restarts

[edit | edit source]

We note that ๐ฑ1 is computed by the gradient descent method applied to ๐ฑ0. Setting ฮฒk=0 would similarly make ๐ฑk+1 computed by the gradient descent method from ๐ฑk, i.e., can be used as a simple implementation of a restart of the conjugate gradient iterations.[4] Restarts could slow down convergence, but may improve stability if the conjugate gradient method misbehaves, e.g., due to round-off error.

Explicit residual calculation

[edit | edit source]

The formulas ๐ฑk+1:=๐ฑk+ฮฑk๐ฉk and ๐ซk:=๐›โˆ’๐€๐ฑk, which both hold in exact arithmetic, make the formulas ๐ซk+1:=๐ซkโˆ’ฮฑk๐€๐ฉk and ๐ซk+1:=๐›โˆ’๐€๐ฑk+1 mathematically equivalent. The former is used in the algorithm to avoid an extra multiplication by ๐€ since the vector ๐€๐ฉk is already computed to evaluate ฮฑk. The latter may be more accurate, substituting the explicit calculation ๐ซk+1:=๐›โˆ’๐€๐ฑk+1 for the implicit one by the recursion subject to round-off error accumulation, and is thus recommended for an occasional evaluation.[7]

A norm of the residual is typically used for stopping criteria. The norm of the explicit residual ๐ซk+1:=๐›โˆ’๐€๐ฑk+1 provides a guaranteed level of accuracy both in exact arithmetic and in the presence of the rounding errors, where convergence naturally stagnates. In contrast, the implicit residual ๐ซk+1:=๐ซkโˆ’ฮฑk๐€๐ฉk is known to keep getting smaller in amplitude well below the level of rounding errors and thus cannot be used to determine the stagnation of convergence.

Computation of alpha and beta

[edit | edit source]

In the algorithm, ฮฑk is chosen such that ๐ซk+1 is orthogonal to ๐ซk. The denominator is simplified from

ฮฑk=๐ซk๐–ณ๐ซk๐ซk๐–ณ๐€๐ฉk=๐ซk๐–ณ๐ซk๐ฉk๐–ณ๐€๐ฉk

since ๐ซk+1=๐ฉk+1โˆ’๐œทk๐ฉk. The ฮฒk is chosen such that ๐ฉk+1 is conjugate to ๐ฉk. Initially, ฮฒk is

ฮฒk=โˆ’๐ซk+1๐–ณ๐€๐ฉk๐ฉk๐–ณ๐€๐ฉk

using

๐ซk+1=๐ซkโˆ’ฮฑk๐€๐ฉk

and equivalently

๐€๐ฉk=1ฮฑk(๐ซkโˆ’๐ซk+1),

the numerator of ฮฒk is rewritten as

๐ซk+1๐–ณ๐€๐ฉk=1ฮฑk๐ซk+1๐–ณ(๐ซkโˆ’๐ซk+1)=โˆ’1ฮฑk๐ซk+1๐–ณ๐ซk+1

because ๐ซk+1 and ๐ซk are orthogonal by design. The denominator is rewritten as

๐ฉk๐–ณ๐€๐ฉk=(๐ซk+ฮฒkโˆ’1๐ฉkโˆ’1)๐–ณ๐€๐ฉk=1ฮฑk๐ซk๐–ณ(๐ซkโˆ’๐ซk+1)=1ฮฑk๐ซk๐–ณ๐ซk

using that the search directions ๐ฉk are conjugated and again that the residuals are orthogonal. This gives the ฮฒ in the algorithm after cancelling ฮฑk.

using LinearAlgebra

"""
    x = conjugate_gradient(A, b, x0 = zero(b); atol=length(b)*eps(norm(b))

Return the solution to `A * x = b` using the conjugate gradient method.
`A` must be a positive definite matrix or other linear operator.
`x0` is the initial guess for the solution (default is the zero vector).
`atol` is the absolute tolerance on the magnitude of the residual `b - A * x`
for convergence (default is machine epsilon).

Returns the approximate solution vector `x`.
"""
function conjugate_gradient(
    A, b::AbstractVector, x0::AbstractVector = zero(b); atol=length(b)*eps(norm(b))
)
    x = copy(x0)                        # initialize the solution
    r = b - A * x0                      # initial residual
    p = copy(r)                         # initial search direction
    rยฒold = r' * r                      # squared norm of residual

    k = 0
    while rยฒold > atol^2                # iterate until convergence
        Ap = A * p                      # search direction
        ฮฑ = rยฒold / (p' * Ap)           # step size
        @. x += ฮฑ * p                   # update solution
        # Update residual:
        if (k + 1) % 16 == 0            # every 16 iterations, recompute residual from scratch 
            r .= b .- A * x             # to avoid accumulation of numerical errors
        else
            @. r -= ฮฑ * Ap              # use the updating formula that saves one matrix-vector product
        end
        rยฒnew = r' * r
        @. p = r + (rยฒnew / rยฒold) * p  # update search direction
        rยฒold = rยฒnew                   # update squared residual norm
        k += 1
    end

    return x
end

Example code in MATLAB

[edit | edit source]
function x = conjugate_gradient(A, b, x0, tol)
% Return the solution to `A * x = b` using the conjugate gradient method.
% Reminder: A should be symmetric and positive definite.

    if nargin < 4
        tol = eps;
    end

    r = b - A * x0;
    p = r;
    rsold = r' * r;

    x = x0;

    while sqrt(rsold) > tol
        Ap = A * p;
        alpha = rsold / (p' * Ap);
        x = x + alpha * p;
        r = r - alpha * Ap;
        rsnew = r' * r;
        p = r + (rsnew / rsold) * p;
        rsold = rsnew;
    end
end


Numerical example

[edit | edit source]

Consider the linear system Ax = b given by

๐€๐ฑ=[4113][x1x2]=[12],

we will perform two steps of the conjugate gradient method beginning with the initial guess

๐ฑ0=[21]

in order to find an approximate solution to the system.

Solution

[edit | edit source]

For reference, the exact solution is

๐ฑ=[111711]โ‰ˆ[0.09090.6364]

Our first step is to calculate the residual vector r0 associated with x0. This residual is computed from the formula r0 = b - Ax0, and in our case is equal to

๐ซ0=[12]โˆ’[4113][21]=[โˆ’8โˆ’3]=๐ฉ0.

Since this is the first iteration, we will use the residual vector r0 as our initial search direction p0; the method of selecting pk will change in further iterations.

We now compute the scalar ฮฑ0 using the relationship

ฮฑ0=๐ซ0๐–ณ๐ซ0๐ฉ0๐–ณ๐€๐ฉ0=[โˆ’8โˆ’3][โˆ’8โˆ’3][โˆ’8โˆ’3][4113][โˆ’8โˆ’3]=73331โ‰ˆ0.2205

We can now compute x1 using the formula

๐ฑ1=๐ฑ0+ฮฑ0๐ฉ0=[21]+73331[โˆ’8โˆ’3]โ‰ˆ[0.23560.3384].

This result completes the first iteration, the result being an "improved" approximate solution to the system, x1. We may now move on and compute the next residual vector r1 using the formula

๐ซ1=๐ซ0โˆ’ฮฑ0๐€๐ฉ0=[โˆ’8โˆ’3]โˆ’73331[4113][โˆ’8โˆ’3]โ‰ˆ[โˆ’0.28100.7492].

Our next step in the process is to compute the scalar ฮฒ0 that will eventually be used to determine the next search direction p1.

ฮฒ0=๐ซ1๐–ณ๐ซ1๐ซ0๐–ณ๐ซ0โ‰ˆ[โˆ’0.28100.7492][โˆ’0.28100.7492][โˆ’8โˆ’3][โˆ’8โˆ’3]=0.0088.

Now, using this scalar ฮฒ0, we can compute the next search direction p1 using the relationship

๐ฉ1=๐ซ1+ฮฒ0๐ฉ0โ‰ˆ[โˆ’0.28100.7492]+0.0088[โˆ’8โˆ’3]=[โˆ’0.35110.7229].

We now compute the scalar ฮฑ1 using our newly acquired p1 using the same method as that used for ฮฑ0.

ฮฑ1=๐ซ1๐–ณ๐ซ1๐ฉ1๐–ณ๐€๐ฉ1โ‰ˆ[โˆ’0.28100.7492][โˆ’0.28100.7492][โˆ’0.35110.7229][4113][โˆ’0.35110.7229]=0.4122.

Finally, we find x2 using the same method as that used to find x1.

๐ฑ2=๐ฑ1+ฮฑ1๐ฉ1โ‰ˆ[0.23560.3384]+0.4122[โˆ’0.35110.7229]=[0.09090.6364].

The result, x2, is a "better" approximation to the system's solution than x1 and x0. If exact arithmetic were to be used in this example instead of limited-precision, then the exact solution would theoretically have been reached after n = 2 iterations (n being the order of the system).

Finite Termination Property

[edit | edit source]

Under exact arithmetic, the number of iterations required is no more than the order of the matrix. This behavior is known as the finite termination property of the conjugate gradient method. It refers to the method's ability to reach the exact solution of a linear system in a finite number of stepsโ€”at most equal to the dimension of the systemโ€”when exact arithmetic is used. This property arises from the fact that, at each iteration, the method generates a residual vector that is orthogonal to all previous residuals. These residuals form a mutually orthogonal set.

In an n-dimensional space, it is impossible to construct more than n linearly independent and mutually orthogonal vectors unless one of them is the zero vector. Therefore, once a zero residual appears, the method has reached the solution and must terminate. This ensures that the conjugate gradient method converges in at most n steps.

To demonstrate this, consider the system:

A=[3โˆ’2โˆ’24],๐›=[11]

We start from an initial guess ๐ฑ0=[12]. Since A is symmetric positive-definite and the system is 2-dimensional, the conjugate gradient method should find the exact solution in no more than 2 steps. The following MATLAB code demonstrates this behavior:

A = [3, -2; -2, 4];
x_true = [1; 1];
b = A * x_true;

x = [1; 2];             % initial guess
r = b - A * x;
p = r;

for k = 1:2
    Ap = A * p;
    alpha = (r' * r) / (p' * Ap);
    x = x + alpha * p;
    r_new = r - alpha * Ap;
    beta = (r_new' * r_new) / (r' * r);
    p = r_new + beta * p;
    r = r_new;
end

disp('Exact solution:');
disp(x);

The output confirms that the method reaches [11] after two iterations, consistent with the theoretical prediction. This example illustrates how the conjugate gradient method behaves as a direct method under idealized conditions.

Application to Sparse Systems

[edit | edit source]

The finite termination property also has practical implications in solving large sparse systems, which frequently arise in scientific and engineering applications. For instance, discretizing the two-dimensional Laplace equation โˆ‡2u=0 using finite differences on a uniform grid leads to a sparse linear system A๐ฑ=๐›, where A is symmetric and positive definite.

Using a 5ร—5 interior grid yields a 25ร—25 system, and the coefficient matrix A has a five-point stencil pattern. Each row of A contains at most five nonzero entries corresponding to the central point and its immediate neighbors. For example, the matrix generated from such a grid may look like:

A=[4โˆ’10โ‹ฏโˆ’10โ‹ฏโˆ’14โˆ’1โ‹ฏ00โ‹ฏ0โˆ’14โˆ’100โ‹ฏโ‹ฎโ‹ฎโ‹ฑโ‹ฑโ‹ฑโ‹ฎโˆ’10โ‹ฏโˆ’14โˆ’1โ‹ฏ00โ‹ฏ0โˆ’14โ‹ฏโ‹ฎโ‹ฎโ‹ฏโ‹ฏโ‹ฏโ‹ฑ]

Although the system dimension is 25, the conjugate gradient method is theoretically guaranteed to terminate in at most 25 iterations under exact arithmetic. In practice, convergence often occurs in far fewer steps due to the matrix's spectral properties. This efficiency makes CGM particularly attractive for solving large-scale systems arising from partial differential equations, such as those found in heat conduction, fluid dynamics, and electrostatics.

Convergence properties

[edit | edit source]

The conjugate gradient method can theoretically be viewed as a direct method, as in the absence of round-off error it produces the exact solution after a finite number of iterations, which is not larger than the size of the matrix. In practice, the exact solution is never obtained since the conjugate gradient method is unstable with respect to even small perturbations, e.g., most directions are not in practice conjugate, due to a degenerative nature of generating the Krylov subspaces.

As an iterative method, the conjugate gradient method monotonically (in the energy norm) improves approximations ๐ฑk to the exact solution and may reach the required tolerance after a relatively small (compared to the problem size) number of iterations. The improvement is typically linear and its speed is determined by the condition number ฮบ(A) of the system matrix A: the larger ฮบ(A) is, the slower the improvement.[8]

However, an interesting case appears when the eigenvalues are spaced logarithmically for a large symmetric matrix. For example, let A=QDQT where Q is a random orthogonal matrix and D is a diagonal matrix with eigenvalues ranging from ฮปn=1 to ฮป1=106, spaced logarithmically. Despite the finite termination property of CGM, where the exact solution should theoretically be reached in at most n steps, the method may exhibit stagnation in convergence. In such a scenario, even after many more iterationsโ€”e.g., ten times the matrix sizeโ€”the error may only decrease modestly (e.g., to 10โˆ’5). Moreover, the iterative error may oscillate significantly, making it unreliable as a stopping condition. This poor convergence is not explained by the condition number alone (e.g., ฮบ2(A)=106), but rather by the eigenvalue distribution itself. When the eigenvalues are more evenly spaced or randomly distributed, such convergence issues are typically absent, highlighting that CGM performance depends not only on ฮบ(A) but also on how the eigenvalues are distributed.[9]

If ฮบ(A) is large, preconditioning is commonly used to replace the original system ๐€๐ฑโˆ’๐›=0 with ๐Œโˆ’1(๐€๐ฑโˆ’๐›)=0 such that ฮบ(๐Œโˆ’1๐€) is smaller than ฮบ(๐€), see below.

Convergence theorem

[edit | edit source]

Define a subset of polynomials as

ฮ k*:={ pโˆˆฮ k : p(0)=1 },

where ฮ k is the set of polynomials of maximal degree k.

Let (๐ฑk)k be the iterative approximations of the exact solution ๐ฑ*, and define the errors as ๐žk:=๐ฑkโˆ’๐ฑ*. Now, the rate of convergence can be approximated as [4][10]

โ€–๐žkโ€–๐€=minpโˆˆฮ k*โ€–p(๐€)๐ž0โ€–๐€โ‰คminpโˆˆฮ k*maxฮปโˆˆฯƒ(๐€)|p(ฮป)| โ€–๐ž0โ€–๐€โ‰ค2(ฮบ(๐€)โˆ’1ฮบ(๐€)+1)k โ€–๐ž0โ€–๐€โ‰ค2exp(โˆ’2kฮบ(๐€)) โ€–๐ž0โ€–๐€,

where ฯƒ(๐€) denotes the spectrum, and ฮบ(๐€) denotes the condition number.

This shows k=12ฮบ(๐€)log(โ€–๐ž0โ€–๐€ฮตโˆ’1) iterations suffices to reduce the error to 2ฮต for any ฮต>0.

Note, the important limit when ฮบ(๐€) tends to โˆž

ฮบ(๐€)โˆ’1ฮบ(๐€)+1โ‰ˆ1โˆ’2ฮบ(๐€)forฮบ(๐€)โ‰ซ1.

This limit shows a faster convergence rate compared to the iterative methods of Jacobi or Gaussโ€“Seidel which scale as โ‰ˆ1โˆ’2ฮบ(๐€).

No round-off error is assumed in the convergence theorem, but the convergence bound is commonly valid in practice as theoretically explained[5] by Anne Greenbaum.

Practical convergence

[edit | edit source]

If initialized randomly, the first stage of iterations is often the fastest, as the error is eliminated within the Krylov subspace that initially reflects a smaller effective condition number. The second stage of convergence is typically well defined by the theoretical convergence bound with ฮบ(๐€), but may be super-linear, depending on a distribution of the spectrum of the matrix A and the spectral distribution of the error.[5] In the last stage, the smallest attainable accuracy is reached and the convergence stalls or the method may even start diverging. In typical scientific computing applications in double-precision floating-point format for matrices of large sizes, the conjugate gradient method uses a stopping criterion with a tolerance that terminates the iterations during the first or second stage.

The preconditioned conjugate gradient method

[edit | edit source]

In most cases, preconditioning is necessary to ensure fast convergence of the conjugate gradient method. If ๐Œโˆ’1 is symmetric positive-definite and ๐Œโˆ’1๐€ has a better condition number than ๐€, a preconditioned conjugate gradient method can be used. It takes the following form:[11]

๐ซ0:=๐›โˆ’๐€๐ฑ0
Solve:๐Œ๐ณ0:=๐ซ0
๐ฉ0:=๐ณ0
k:=0
repeat
ฮฑk:=๐ซk๐–ณ๐ณk๐ฉk๐–ณ๐€๐ฉk
๐ฑk+1:=๐ฑk+ฮฑk๐ฉk
๐ซk+1:=๐ซkโˆ’ฮฑk๐€๐ฉk
if rk+1 is sufficiently small then exit loop end if
Solve ๐Œ๐ณk+1:=๐ซk+1
ฮฒk:=๐ซk+1๐–ณ๐ณk+1๐ซk๐–ณ๐ณk
๐ฉk+1:=๐ณk+1+ฮฒk๐ฉk
k:=k+1
end repeat
The result is xk+1

The above formulation is equivalent to applying the regular conjugate gradient method to the preconditioned system[12]

๐„โˆ’1๐€(๐„โˆ’1)๐–ณ๐ฑ^=๐„โˆ’1๐›

where

๐„๐„๐–ณ=๐Œ,๐ฑ^=๐„๐–ณ๐ฑ.

The Cholesky decomposition of the preconditioner must be used to keep the symmetry (and positive definiteness) of the system. However, this decomposition does not need to be computed, and it is sufficient to know ๐Œโˆ’1. It can be shown that ๐„โˆ’1๐€(๐„โˆ’1)๐–ณ has the same spectrum as ๐Œโˆ’1๐€.

The preconditioner matrix ๐Œ has to be symmetric positive-definite and fixed, i.e., cannot change from iteration to iteration. If any of these assumptions on the preconditioner is violated, the behavior of the preconditioned conjugate gradient method may become unpredictable.

An example of a commonly used preconditioner is the incomplete Cholesky factorization.[13]

Using the preconditioner in practice

[edit | edit source]

It is important to keep in mind that we don't want to invert the matrix ๐Œ explicitly in order to get ๐Œโˆ’1 for use in the process, since inverting ๐Œ would take more time/computational resources than solving the conjugate gradient algorithm itself. As an example, let's say that we are using a preconditioner coming from incomplete Cholesky factorization. The resulting matrix is the lower triangular matrix ๐‹, and the preconditioner matrix is:

๐Œ=๐‹๐‹๐–ณ

Then we have to solve:

๐Œ๐ณ=๐ซ

๐ณ=๐Œโˆ’1๐ซ

But:

๐Œโˆ’1=(๐‹โˆ’1)๐–ณ๐‹โˆ’1

Then:

๐ณ=(๐‹โˆ’1)๐–ณ๐‹โˆ’1๐ซ

Let's take an intermediary vector ๐š:

๐š=๐‹โˆ’1๐ซ

๐ซ=๐‹๐š

Since ๐ซ and ๐‹ and known, and ๐‹ is lower triangular, solving for ๐š is easy and computationally cheap by using forward substitution. Then, we substitute ๐š in the original equation:

๐ณ=(๐‹โˆ’1)๐–ณ๐š

๐š=๐‹๐–ณ๐ณ

Since ๐š and ๐‹๐–ณ are known, and ๐‹๐–ณ is upper triangular, solving for ๐ณ is easy and computationally cheap by using backward substitution.

Using this method, there is no need to invert ๐Œ or ๐‹ explicitly at all, and we still obtain ๐ณ.

The flexible preconditioned conjugate gradient method

[edit | edit source]

In numerically challenging applications, sophisticated preconditioners are used, which may lead to variable preconditioning, changing between iterations. Even if the preconditioner is symmetric positive-definite on every iteration, the fact that it may change makes the arguments above invalid, and in practical tests leads to a significant slow down of the convergence of the algorithm presented above. Using the Polakโ€“Ribiรจre formula

ฮฒk:=๐ซk+1๐–ณ(๐ณk+1โˆ’๐ณk)๐ซk๐–ณ๐ณk

instead of the Fletcherโ€“Reeves formula

ฮฒk:=๐ซk+1๐–ณ๐ณk+1๐ซk๐–ณ๐ณk

may dramatically improve the convergence in this case.[14] This version of the preconditioned conjugate gradient method can be called[15] flexible, as it allows for variable preconditioning. The flexible version is also shown[16] to be robust even if the preconditioner is not symmetric positive definite (SPD).

The implementation of the flexible version requires storing an extra vector. For a fixed SPD preconditioner, ๐ซk+1๐–ณ๐ณk=0, so both formulas for ฮฒk are equivalent in exact arithmetic, i.e., without the round-off error.

The mathematical explanation of the better convergence behavior of the method with the Polakโ€“Ribiรจre formula is that the method is locally optimal in this case, in particular, it does not converge slower than the locally optimal steepest descent method.[17]

Vs. the locally optimal steepest descent method

[edit | edit source]

In both the original and the preconditioned conjugate gradient methods one only needs to set ฮฒk:=0 in order to make them locally optimal, using the line search, steepest descent methods. With this substitution, vectors p are always the same as vectors z, so there is no need to store vectors p. Thus, every iteration of these steepest descent methods is a bit cheaper compared to that for the conjugate gradient methods. However, the latter converge faster, unless a (highly) variable and/or non-SPD preconditioner is used, see above.

Conjugate gradient method as optimal feedback controller for double integrator

[edit | edit source]

The conjugate gradient method can also be derived using optimal control theory.[18] In this approach, the conjugate gradient method falls out as an optimal feedback controller,u=k(x,v):=โˆ’ฮณaโˆ‡f(x)โˆ’ฮณbv for the double integrator system,xห™=v,vห™=u The quantities ฮณa and ฮณb are variable feedback gains.[18]

Conjugate gradient on the normal equations

[edit | edit source]

The conjugate gradient method can be applied to an arbitrary n-by-m matrix by applying it to normal equations ATA and right-hand side vector ATb, since ATA is a symmetric positive-semidefinite matrix for any A. The result is conjugate gradient on the normal equations (CGN or CGNR).

ATAx = ATb

As an iterative method, it is not necessary to form ATA explicitly in memory but only to perform the matrixโ€“vector and transpose matrixโ€“vector multiplications. Therefore, CGNR is particularly useful when A is a sparse matrix since these operations are usually extremely efficient. However the downside of forming the normal equations is that the condition number ฮบ(ATA) is equal to ฮบ2(A) and so the rate of convergence of CGNR may be slow and the quality of the approximate solution may be sensitive to roundoff errors. Finding a good preconditioner is often an important part of using the CGNR method.

Several algorithms have been proposed (e.g., CGLS, LSQR). The LSQR algorithm purportedly has the best numerical stability when A is ill-conditioned, i.e., A has a large condition number.

Conjugate gradient method for complex Hermitian matrices

[edit | edit source]

The conjugate gradient method with a trivial modification is extendable to solving, given complex-valued matrix A and vector b, the system of linear equations ๐€๐ฑ=๐› for the complex-valued vector x, where A is Hermitian (i.e., A' = A) and positive-definite matrix, and the symbol ' denotes the conjugate transpose. The trivial modification is simply substituting the conjugate transpose for the real transpose everywhere.

Advantages and disadvantages

[edit | edit source]

The advantages and disadvantages of the conjugate gradient methods are summarized in the lecture notes by Nemirovsky and BenTal.[19]: Sec.7.3 

A pathological example

[edit | edit source]

This example is from [20] Let tโˆˆ(0,1), and defineW=[ttt1+ttt1+tttโ‹ฑโ‹ฑโ‹ฑtt1+t],b=[10โ‹ฎ0]Since W is invertible, there exists a unique solution to Wx=b. Solving it by conjugate gradient descent gives us rather bad convergence:โ€–bโˆ’Wxkโ€–2=(1/t)k,โ€–bโˆ’Wxnโ€–2=0In words, during the CG process, the error grows exponentially, until it suddenly becomes zero as the unique solution is found.

See also

[edit | edit source]

References

[edit | edit source]
  1. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  2. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  3. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  4. ^ a b c d Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  5. ^ a b c Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  6. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  7. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  8. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  9. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  10. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  11. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  12. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  13. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  14. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  15. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  16. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  17. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  18. ^ a b Ross, I. M., "An Optimal Control Theory for Accelerated Optimization," Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value)., 2019.
  19. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  20. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).

Further reading

[edit | edit source]
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Gรฉrard Meurant: "Detection and correction of silent errors in the conjugate gradient algorithm", Numerical Algorithms, vol.92 (2023), pp.869-891. url=https://doi.org/10.1007/s11075-022-01380-1
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
[edit | edit source]
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).

Lua error in Module:Authority_control at line 153: attempt to index field 'wikibase' (a nil value).