Rayleigh–Ritz method

The Rayleigh–Ritz method is a direct numerical method of approximating eigenvalues, originated in the context of solving physical boundary value problems and named after Lord Rayleigh and Walther Ritz.

In this method, an infinite-dimensional linear operator is approximated by a finite-dimensional compression, on which we can use an eigenvalue algorithm.

It is used in all applications that involve approximating eigenvalues and eigenvectors, often under different names. In quantum mechanics, where a system of particles is described using a Hamiltonian, the Ritz method uses trial wave functions to approximate the ground state eigenfunction with the lowest energy. In the finite element method context, mathematically the same algorithm is commonly called the Ritz-Galerkin method. The Rayleigh–Ritz method or Ritz method terminology is typical in mechanical and structural engineering to approximate the eigenmodes and resonant frequencies of a structure.

Naming and attribution edit

The name of the method and its origin story have been debated by histroians.[1][2] It has been called Ritz method after Walther Ritz, since the numerical procedure has been published by Walther Ritz in 1908-1909. According to A. W. Leissa,[1] Lord Rayleigh wrote a paper congratulating Ritz on his work in 1911, but stating that he himself had used Ritz's method in many places in his book and in another publication. This statement, although later disputed, and the fact that the method in the trivial case of a single vector results in the Rayleigh quotient make the case for the name Rayleigh–Ritz method. According to S. Ilanko,[2] citing Richard Courant, both Lord Rayleigh and Walther Ritz independently conceived the idea of utilizing the equivalence between boundary value problems of partial differential equations on the one hand and problems of the calculus of variations on the other hand for numerical calculation of the solutions, by substituting for the variational problems simpler approximating extremum problems in which a finite number of parameters need to be determined; see the article Ritz method for details. Ironically for the debate, the modern justification of the algorithm drops the calculus of variations in favor of the simpler and more general approach of orthogonal projection as in Galerkin method named after Boris Galerkin, thus leading also to the Ritz-Galerkin method naming.[citation needed]

Method edit

Let   be a linear operator on a Hilbert space  , with inner product  . Now consider a finite set of functions  . Depending on the application these functions may be:

One could use the orthonormal basis generated from the eigenfunctions of the operator, which will produce diagonal approximating matrices, but in this case we would have already had to calculate the spectrum.

We now approximate   by  , which is defined as the matrix with entries[3]

 

and solve the eigenvalue problem  . It can be shown that the matrix   is the compression of   to  .[3]

For differential operators (such as Sturm-Liouville operators), the inner product   can be replaced by the weak formulation  .[4][6]

If a subset of the orthonormal basis was used to find the matrix, the eigenvectors of   will be linear combinations of orthonormal basis functions, and as a result they will be approximations of the eigenvectors of  .[7]

Properties edit

Spectral pollution edit

It is possible for the Rayleigh–Ritz method to produce values which do not converge to actual values in the spectrum of the operator as the truncation gets large. These values are known as spectral pollution.[3][5][8] In some cases (such as for the Schrödinger equation), there is no approximation which both includes all eigenvalues of the equation, and contains no pollution.[9]

The spectrum of the compression (and thus pollution) is bounded by the numerical range of the operator; in many cases it is bounded by a subset of the numerical range known as the essential numerical range.[10][11]

For matrix eigenvalue problems edit

In numerical linear algebra, the Rayleigh–Ritz method is commonly[12] applied to approximate an eigenvalue problem

 
for the matrix   of size   using a projected matrix of a smaller size  , generated from a given matrix   with orthonormal columns. The matrix version of the algorithm is the most simple:
  1. Compute the   matrix  , where   denotes the complex-conjugate transpose of  
  2. Solve the eigenvalue problem  
  3. Compute the Ritz vectors   and the Ritz value  
  4. Output approximations  , called the Ritz pairs, to eigenvalues and eigenvectors of the original matrix  .

If the subspace with the orthonormal basis given by the columns of the matrix   contains   vectors that are close to eigenvectors of the matrix  , the Rayleigh–Ritz method above finds   Ritz vectors that well approximate these eigenvectors. The easily computable quantity   determines the accuracy of such an approximation for every Ritz pair.

In the easiest case  , the   matrix   turns into a unit column-vector  , the   matrix   is a scalar that is equal to the Rayleigh quotient  , the only   solution to the eigenvalue problem is   and  , and the only one Ritz vector is   itself. Thus, the Rayleigh–Ritz method turns into computing of the Rayleigh quotient if  .

Another useful connection to the Rayleigh quotient is that   for every Ritz pair  , allowing to derive some properties of Ritz values   from the corresponding theory for the Rayleigh quotient. For example, if   is a Hermitian matrix, its Rayleigh quotient (and thus its every Ritz value) is real and takes values within the closed interval of the smallest and largest eigenvalues of  .

Example edit

The matrix

 
has eigenvalues   and the corresponding eigenvectors
 
Let us take
 
then
 
with eigenvalues   and the corresponding eigenvectors
 
so that the Ritz values are   and the Ritz vectors are
 
We observe that each one of the Ritz vectors is exactly one of the eigenvectors of   for the given   as well as the Ritz values give exactly two of the three eigenvalues of  . A mathematical explanation for the exact approximation is based on the fact that the column space of the matrix   happens to be exactly the same as the subspace spanned by the two eigenvectors   and   in this example.

For matrix singular value problems edit

Truncated singular value decomposition (SVD) in numerical linear algebra can also use the Rayleigh–Ritz method to find approximations to left and right singular vectors of the matrix   of size   in given subspaces by turning the singular value problem into an eigenvalue problem.

Using the normal matrix edit

The definition of the singular value   and the corresponding left and right singular vectors is   and  . Having found one set (left of right) of approximate singular vectors and singular values by applying naively the Rayleigh–Ritz method to the Hermitian normal matrix   or  , whichever one is smaller size, one could determine the other set of left of right singular vectors simply by dividing by the singular values, i.e.,   and  . However, the division is unstable or fails for small or zero singular values.

An alternative approach, e.g., defining the normal matrix as   of size  , takes advantage of the fact that for a given   matrix   with orthonormal columns the eigenvalue problem of the Rayleigh–Ritz method for the   matrix

 
can be interpreted as a singular value problem for the   matrix  . This interpretation allows simple simultaneous calculation of both left and right approximate singular vectors as follows.
  1. Compute the   matrix  .
  2. Compute the thin, or economy-sized, SVD   with   matrix  ,   diagonal matrix  , and   matrix  .
  3. Compute the matrices of the Ritz left   and right   singular vectors.
  4. Output approximations  , called the Ritz singular triplets, to selected singular values and the corresponding left and right singular vectors of the original matrix   representing an approximate Truncated singular value decomposition (SVD) with left singular vectors restricted to the column-space of the matrix  .

The algorithm can be used as a post-processing step where the matrix   is an output of an eigenvalue solver, e.g., such as LOBPCG, approximating numerically selected eigenvectors of the normal matrix  .

Example edit

The matrix

 
has its normal matrix
 
singular values   and the corresponding thin SVD
 
where the columns of the first multiplier from the complete set of the left singular vectors of the matrix  , the diagonal entries of the middle term are the singular values, and the columns of the last multiplier transposed (although the transposition does not change it)
 
are the corresponding right singular vectors.

Let us take

 
with the column-space that is spanned by the two exact right singular vectors
 
corresponding to the singular values 1 and 2.

Following the algorithm step 1, we compute

 
and on step 2 its thin SVD   with
 
Thus we already obtain the singular values 2 and 1 from   and from   the corresponding two left singular vectors   as   and  , which span the column-space of the matrix  , explaining why the approximations are exact for the given  .

Finally, step 3 computes the matrix  

 
recovering from its rows the two right singular vectors   as   and  . We validate the first vector:  
 
and  
 
Thus, for the given matrix   with its column-space that is spanned by two exact right singular vectors, we determine these right singular vectors, as well as the corresponding left singular vectors and the singular values, all exactly. For an arbitrary matrix  , we obtain approximate singular triplets which are optimal given   in the sense of optimality of the Rayleigh–Ritz method.

Applications and examples edit

In quantum physics edit

In quantum physics, where the spectrum of the Hamiltonian is the set of discrete energy levels allowed by a quantum mechanical system, the Rayleigh–Ritz method is used to approximate the energy states and wavefunctions of a complicated atomic or nuclear system.[7]

In dynamical systems edit

The Koopman operator allows a finite-dimensional nonlinear system to be encoded as an infinite-dimensional linear system. In general, both of these problems are difficult to solve, but for the latter we can use the Ritz-Galerkin method to approximate a solution.[13]

See also edit

Notes and references edit

  1. ^ a b Leissa, A.W. (2005). "The historical bases of the Rayleigh and Ritz methods". Journal of Sound and Vibration. 287 (4–5): 961–978. Bibcode:2005JSV...287..961L. doi:10.1016/j.jsv.2004.12.021.
  2. ^ a b Ilanko, Sinniah (2009). "Comments on the historical bases of the Rayleigh and Ritz methods". Journal of Sound and Vibration. 319 (1–2): 731–733. Bibcode:2009JSV...319..731I. doi:10.1016/j.jsv.2008.06.001.
  3. ^ a b c d Davies, E. B.; Plum, M. (2003). "Spectral Pollution". IMA Journal of Numerical Analysis.
  4. ^ a b Süli, Endre; Mayers, David (2003). An Introduction to Numerical Analysis. Cambridge University Press. ISBN 0521007941.
  5. ^ a b Levitin, Michael; Shargorodsky, Eugene (2004). "Spectral pollution and second order relative spectra for self-adjoint operators". IMA Journal of Numerical Analysis.
  6. ^ Pryce, John D. (1994). Numerical Solution of Sturm-Liouville Problems. Oxford University Press. ISBN 0198534159.
  7. ^ a b Arfken, George B.; Weber, Hans J. (2005). Mathematical Methods For Physicists (6th ed.). Academic Press.
  8. ^ Colbrook, Matthew. "Unscrambling the Infinite: Can we Compute Spectra?". Mathematics Today. Institute of Mathematics and its Applications.
  9. ^ Colbrook, Matthew; Roman, Bogdan; Hansen, Anders (2019). "How to Compute Spectra with Error Control". Physical Review Letters.
  10. ^ Pokrzywa, Andrzej (1979). "Method of orthogonal projections and approximation of the spectrum of a bounded operator". Studia Mathematica.
  11. ^ Bögli, Sabine; Marletta, Marco; Tretter, Christiane (2020). "The essential numerical range for unbounded linear operators". Journal of Functional Analysis.
  12. ^ Trefethen, Lloyd N.; Bau, III, David (1997). Numerical Linear Algebra. SIAM. p. 254. ISBN 978-0-89871-957-4.
  13. ^ Servadio, Simone; Arnas, David; Linares, Richard. "A Koopman Operator Tutorial with Orthogonal Polynomials". arXiv.

External links edit