Skip to article frontmatterSkip to article content

Inverse iteration

Power iteration finds only the dominant eigenvalue. We next show that it can be adapted to find any eigenvalue, provided you start with a reasonably good estimate of it. Some simple linear algebra is all that is needed.

Consider first part 2 of the theorem with s=0s=0, and suppose that A\mathbf{A} has a smallest eigenvalue,

λnλn1>λ1.|\lambda_n| \ge |\lambda_{n-1}| \ge \cdots > |\lambda_1|.

Then clearly

λ11>λ21λn1,|\lambda_1^{-1}| > |\lambda_{2}^{-1}| \ge \cdots \ge |\lambda_n^{-1}|,

and A1\mathbf{A}^{-1} has a dominant eigenvalue. Hence power iteration on A1\mathbf{A}^{-1} can be used to find the eigenvalue of A\mathbf{A} closest to zero. For nonzero values of ss, then we suppose there is an ordering

λnsλ2s>λ1s.|\lambda_n-s| \ge \cdots \ge |\lambda_2-s| > |\lambda_1-s|.

Then it follows that

λ1s1>λ2s1λns1,|\lambda_1-s|^{-1} > |\lambda_{2}-s|^{-1} \ge \cdots \ge |\lambda_n-s|^{-1},

and power iteration on the matrix (AsI)1(\mathbf{A}-s\mathbf{I})^{-1} converges to (λ1s)1(\lambda_1-s)^{-1}, which is easily solved for λ1\lambda_1 itself.

8.3.1Algorithm

A literal application of Algorithm 8.2.1 would include the step

yk=(AsI)1xk.\mathbf{y}_k = (\mathbf{A}-s\mathbf{I})^{-1} \mathbf{x}_k.

As always, we do not want to explicitly find the inverse of a matrix. Instead we should write this step as the solution of a linear system.

Each pass of inverse iteration requires the solution of a linear system of equations with the matrix B=AsI\mathbf{B}=\mathbf{A}-s\mathbf{I}. This solution might use methods we consider later in this chapter. Here, we use (sparse) PLU factorization and hope for the best. Since the matrix B\mathbf{B} is constant, the factorization needs to be done only once for all iterations. The details are in Function 8.3.2.

8.3.2Convergence

The convergence is linear, at a rate found by reinterpreting (8.2.10) with (AsI)1(\mathbf{A}-s\mathbf{I})^{-1} in place of A\mathbf{A}:

βk+1λ1βkλ1λ1sλ2s as k,\frac{\beta_{k+1} - \lambda_1}{\beta_{k} - \lambda_1} \rightarrow \frac{ \lambda_1 - s } {\lambda_2 - s}\quad \text{ as } \quad k\rightarrow \infty,

with the eigenvalues ordered as in (8.3.3). Thus, the convergence is best when the shift ss is close to the target eigenvalue λ1\lambda_1, specifically when it is much closer to that eigenvalue than to any other.

8.3.3Dynamic shifting

There is a clear opportunity for positive feedback in Algorithm 8.3.1. The convergence rate of inverse iteration improves as the shift gets closer to the true eigenvalue—and the algorithm computes improving eigenvalue estimates! If we update the shift to s=βks=\beta_k after each iteration, the convergence accelerates. You are asked to implement this algorithm in Exercise 6.

Let’s analyze the resulting convergence. If the eigenvalues are ordered by distance to ss, then the convergence is linear with rate λ1s/λ2s|\lambda_1-s|/|\lambda_2-s|. As sλ1s\to\lambda_1, the change in the denominator is negligible. So if the error (λ1s)(\lambda_1-s) is ε, then the error in the next estimate is reduced by a factor O(ϵ)O(\epsilon). That is, ε becomes O(ϵ2)O(\epsilon^2), which is quadratic convergence.

There is a price to pay for this improvement. The matrix of the linear system to be solved, (AsI),(\mathbf{A}-s\mathbf{I}), now changes with each iteration. That means that we can no longer do just one LU factorization for the entire iteration. The speedup in convergence usually makes this tradeoff worthwhile, however.

In practice power and inverse iteration are not as effective as the algorithms used by eigs and based on the mathematics described in the rest of this chapter. However, inverse iteration can be useful for turning an eigenvalue estimate into an eigenvector estimate.

8.3.4Exercises