Fast multipole method

The fast multipole method (FMM) is a numerical technique that was developed to speed up the calculation of long-ranged forces in the n-body problem.

It does this by expanding the system Green's function using a multipole expansion, which allows one to group sources that lie close together and treat them as if they are a single source.

[1] The FMM has also been applied in accelerating the iterative solver in the method of moments (MOM) as applied to computational electromagnetics problems,[2] and in particular in computational bioelectromagnetism.

The FMM was first introduced in this manner by Leslie Greengard and Vladimir Rokhlin Jr.[3] and is based on the multipole expansion of the vector Helmholtz equation.

By treating the interactions between far-away basis functions using the FMM, the corresponding matrix elements do not need to be explicitly stored, resulting in a significant reduction in required memory.

If the FMM is then applied in a hierarchical manner, it can improve the complexity of matrix-vector products in an iterative solver from

in finite arithmetic, i.e., given a tolerance

, the matrix-vector product is guaranteed to be within a tolerance

The dependence of the complexity on the tolerance

This has expanded the area of applicability of the MOM to far greater problems than were previously possible.

The FMM, introduced by Rokhlin Jr. and Greengard has been said to be one of the top ten algorithms of the 20th century.

[4] The FMM algorithm reduces the complexity of matrix-vector multiplication involving a certain type of dense matrix which can arise out of many physical systems.

The FMM has also been applied for efficiently treating the Coulomb interaction in the Hartree–Fock method and density functional theory calculations in quantum chemistry.

In its simplest form, the fast multipole method seeks to evaluate the following function: where

are a set of poles and

are the corresponding pole weights on a set of points

This is the one-dimensional form of the problem, but the algorithm can be easily generalized to multiple dimensions and kernels other than

Naively, evaluating

points requires

The crucial observation behind the fast multipole method is that if the distance between

be the Chebyshev nodes of order

be the corresponding Lagrange basis polynomials.

One can show that the interpolating polynomial: converges quickly with polynomial order,

, provided that the pole is far enough away from the region of interpolation,

The speed-up of the fast multipole method derives from this interpolation: provided that all the poles are "far away", we evaluate the sum only on the Chebyshev nodes at a cost of

, and then interpolate it onto all the desired points at a cost of

is the numerical tolerance, the total cost is

To ensure that the poles are indeed well-separated, one recursively subdivides the unit interval such that only

poles end up in each interval.

One then uses the explicit formula within each interval and interpolation for all intervals that are well-separated.

Fast Multipole Method - interpolation of a pole at x=3 with an order-5 Chebyshev polynomial