The fast multipole method (FMM) is a numerical technique that was developed to speed up the calculation of long-ranged forces in the n-body problem. It does this by expanding the system Green's function using a multipole expansion, which allows one to group sources that lie close together and treat them as if they are a single source.[1]

The FMM has also been applied in accelerating the iterative solver in the method of moments (MOM) as applied to computational electromagnetics problems [2], and in particular in computational bioelectromagnetism. The FMM was first introduced in this manner by Leslie Greengard and Vladimir Rokhlin Jr.[3] and is based on the multipole expansion of the vector Helmholtz equation. By treating the interactions between far-away basis functions using the FMM, the corresponding matrix elements do not need to be explicitly stored, resulting in a significant reduction in required memory. If the FMM is then applied in a hierarchical manner, it can improve the complexity of matrix-vector products in an iterative solver from to in finite arithmetic, i.e., given a tolerance , the matrix-vector product is guaranteed to be within a tolerance The dependence of the complexity on the tolerance is , i.e., the complexity of FMM is . This has expanded the area of applicability of the MOM to far greater problems than were previously possible.

The FMM, introduced by Rokhlin Jr. and Greengard has been said to be one of the top ten algorithms of the 20th century.[4] The FMM algorithm reduces the complexity of matrix-vector multiplication involving a certain type of dense matrix which can arise out of many physical systems.

The FMM has also been applied for efficiently treating the Coulomb interaction in the Hartree–Fock method and density functional theory calculations in quantum chemistry.

Sketch of the Algorithm edit

 
Fast Multipole Method - interpolation of a pole at x=3 with an order-5 Chebyshev polynomial

In its simplest form, the fast multipole method seeks to evaluate the following function:

 ,

where   are a set of poles and   are the corresponding pole weights on a set of points   with  . This is the one-dimensional form of the problem, but the algorithm can be easily generalized to multiple dimensions and kernels other than  .

Naively, evaluating   on   points requires   operations. The crucial observation behind the fast multipole method is that if the distance between   and   is large enough, then   is well-approximated by a polynomial. Specifically, let   be the Chebyshev nodes of order   and let   be the corresponding Lagrange basis polynomials. One can show that the interpolating polynomial:

 

converges quickly with polynomial order,  , provided that the pole is far enough away from the region of interpolation,   and  . This is known as the "local expansion".

The speed-up of the fast multipole method derives from this interpolation: provided that all the poles are "far away", we evaluate the sum only on the Chebyshev nodes at a cost of  , and then interpolate it onto all the desired points at a cost of  :

 

Since  , where   is the numerical tolerance, the total cost is  .

To ensure that the poles are indeed well-separated, one recursively subdivides the unit interval such that only   poles end up in each interval. One then uses the explicit formula within each interval and interpolation for all intervals that are well-separated. This does not spoil the scaling, since one needs at most   levels within the given tolerance.

See also edit

References edit

  1. ^ Rokhlin, Vladimir (1985). "Rapid Solution of Integral Equations of Classic Potential Theory." J. Computational Physics Vol. 60, pp. 187–207.
  2. ^ Nader Engheta, William D. Murphy, Vladimir Rokhlin, and Marius Vassiliou (1992), “The Fast Multipole Method for Electromagnetic Scattering Computation,” IEEE Transactions on Antennas and Propagation 40, 634–641.
  3. ^ "The Fast Multipole Method". Archived from the original on 2011-06-03. Retrieved 2010-12-10.
  4. ^ Cipra, Barry Arthur (May 16, 2000). "The Best of the 20th Century: Editors Name Top 10 Algorithms". SIAM News. 33 (4). Society for Industrial and Applied Mathematics: 2. Retrieved February 27, 2019.

External links edit

Free software edit

  • Puma-EM A high performance, parallelized, open source Method of Moments / Multilevel Fast Multipole Method electromagnetics code.
  • KIFMM3d The Kernel-Independent Fast Multipole 3d Method (kifmm3d) is a new FMM implementation which does not require the explicit multipole expansions of the underlying kernel, and it is based on kernel evaluations.
  • PVFMM A optimized parallel implementation of KIFMM for computing potentials from particle and volume sources.
  • FastBEM Free fast multipole boundary element programs for solving 2D/3D potential, elasticity, stokes flow and acoustic problems.
  • FastFieldSolvers maintains the distribution of the tools, called FastHenry and FastCap, developed at M.I.T. for the solution of Maxwell equations and extraction of circuit parasites (inductance and capacitance) using the FMM.
  • ExaFMM ExaFMM is a CPU/GPU capable 3D FMM code for Laplace/Helmholtz kernels that focuses on parallel scalability.
  • ScalFMM Archived 2017-05-02 at the Wayback Machine ScalFMM is a C++ software library developed at Inria Bordeaux with high emphasis on genericity and parallelization (using OpenMP/MPI).
  • DASHMM DASHMM is a C++ Software library developed at Indiana University using Asynchronous Multi-Tasking HPX-5 runtime system. It provides a unified execution on shared and distributed memory computers and provides 3D Laplace, Yukawa, and Helmholtz kernels.
  • RECFMM Adaptive FMM with dynamic parallelism on multicores.