# Error function

Plot of the error function

In mathematics, the error function (also called the Gauss error function) is a special function (non-elementary) of sigmoid shape which occurs in probability, statistics and partial differential equations. It is defined as:[1][2]

$\operatorname{erf}(x) = \frac{2}{\sqrt{\pi}}\int_{0}^x e^{-t^2}\,\mathrm dt.$

The complementary error function, denoted erfc, is defined as

\begin{align} \operatorname{erfc}(x) & = 1-\operatorname{erf}(x) \\ & = \frac{2}{\sqrt{\pi}} \int_x^{\infty} e^{-t^2}\,\mathrm dt. \end{align}

The imaginary error function, denoted erfi, is defined as

$\operatorname{erfi}(z) = -i\,\,\operatorname{erf}(i\,z).$

When the error function is evaluated for arbitrary complex arguments z, the resulting complex error function is usually discussed in scaled form as the Faddeeva function:

$w(z) = e^{-z^2}\operatorname{erfc}(-iz) .$

## The name "error function"

The error function is used in measurement theory (using probability and statistics), and although its use in other branches of mathematics has nothing to do with the characterization of measurement errors, the name has stuck.

The error function is related to the cumulative distribution $\Phi$, the integral of the standard normal distribution, by[2]

$\Phi (x) = \frac{1}{2}+ \frac{1}{2} \operatorname{erf} \left(x/ \sqrt{2}\right).$

The error function, evaluated at $\frac{x}{\sigma \sqrt{2}}$ for positive x values, gives the probability that a measurement, under the influence of normally distributed errors with standard deviation $\sigma$, has a distance less than x from the mean value.[3] This function is used in statistics to predict behavior of any sample with respect to the population mean. This usage is similar to the Q-function, which in fact can be written in terms of the error function.

↑Jump back a section

## Properties

Plots in the complex plane

Integrand exp(−z2)
erf(z)

The property $\operatorname{erf} (-z) = -\operatorname{erf} (z)$ means that the error function is an odd function.

For any complex number z:

$\operatorname{erf} (\overline{z}) = \overline{\operatorname{erf}(z)}$

where $\overline{z}$ is the complex conjugate of z.

The integrand ƒ = exp(−z2) and ƒ = erf(z) are shown in the complex z-plane in figures 2 and 3. Level of Im(ƒ) = 0 is shown with a thick green line. Negative integer values of Im(ƒ) are shown with thick red lines. Positive integer values of Im(f) are shown with thick blue lines. Intermediate levels of Im(ƒ) = constant are shown with thin green lines. Intermediate levels of Re(ƒ) = constant are shown with thin red lines for negative values and with thin blue lines for positive values.

At the real axis, erf(z) approaches unity at z → +∞ and −1 at z → −∞. At the imaginary axis, it tends to ±i∞.

### Taylor series

The error function is an entire function; it has no singularities (except that at infinity) and its Taylor expansion always converges.

The defining integral cannot be evaluated in closed form in terms of elementary functions, but by expanding the integrand ez2 into its Taylor series and integrating term by term, one obtains the error function's Taylor series as:

$\operatorname{erf}(z)= \frac{2}{\sqrt{\pi}}\sum_{n=0}^\infin\frac{(-1)^n z^{2n+1}}{n! (2n+1)} =\frac{2}{\sqrt{\pi}} \left(z-\frac{z^3}{3}+\frac{z^5}{10}-\frac{z^7}{42}+\frac{z^9}{216}-\ \cdots\right)$

which holds for every complex number z. The denominator terms are sequence A007680 in the OEIS.

For iterative calculation of the above series, the following alternative formulation may be useful:

$\operatorname{erf}(z)= \frac{2}{\sqrt{\pi}}\sum_{n=0}^\infin\left(z \prod_{k=1}^n {\frac{-(2k-1) z^2}{k (2k+1)}}\right) = \frac{2}{\sqrt{\pi}} \sum_{n=0}^\infin \frac{z}{2n+1} \prod_{k=1}^n \frac{-z^2}{k}$

because $\frac{-(2k-1) z^2}{k (2k+1)}$ expresses the multiplier to turn the kth term into the (k + 1)th term (considering z as the first term).

The error function at +∞ is exactly 1 (see Gaussian integral).

The derivative of the error function follows immediately from its definition:

$\frac{\rm d}{{\rm d}z}\,\mathrm{erf}(z)=\frac{2}{\sqrt{\pi}}\,e^{-z^2}.$

An antiderivative of the error function is

$z\,\operatorname{erf}(z) + \frac{e^{-z^2}}{\sqrt{\pi}}.$

### Inverse functions

The inverse error function can be defined in terms of the Maclaurin series

$\operatorname{erf}^{-1}(z)=\sum_{k=0}^\infin\frac{c_k}{2k+1}\left (\frac{\sqrt{\pi}}{2}z\right )^{2k+1}, \,\!$

where c0 = 1 and

$c_k=\sum_{m=0}^{k-1}\frac{c_m c_{k-1-m}}{(m+1)(2m+1)} = \left\{1,1,\frac{7}{6},\frac{127}{90},\frac{4369}{2520},\ldots\right\}.$

So we have the series expansion (note that common factors have been canceled from numerators and denominators):

$\operatorname{erf}^{-1}(z)=\tfrac{1}{2}\sqrt{\pi}\left (z+\frac{\pi}{12}z^3+\frac{7\pi^2}{480}z^5+\frac{127\pi^3}{40320}z^7+\frac{4369\pi^4}{5806080}z^9+\frac{34807\pi^5}{182476800}z^{11}+\cdots\right ).\$

(After cancellation the numerator/denominator fractions are entries A092676/A132467 in the OEIS; without cancellation the numerator terms are given in entry A002067.) Note that the error function's value at ±∞ is equal to ±1.

The inverse complementary error function is defined as

$\operatorname{erfc}^{-1}(1-z) = \operatorname{erf}^{-1}(z).$

### Asymptotic expansion

A useful asymptotic expansion of the complementary error function (and therefore also of the error function) for large real x is

$\mathrm{erfc}(x) = \frac{e^{-x^2}}{x\sqrt{\pi}}\left [1+\sum_{n=1}^\infty (-1)^n \frac{1\cdot3\cdot5\cdots(2n-1)}{(2x^2)^n}\right ]=\frac{e^{-x^2}}{x\sqrt{\pi}}\sum_{n=0}^\infty (-1)^n \frac{(2n-1)!!}{(2x^2)^n},\,$

where (2n – 1)!! is the double factorial: the product of all odd numbers up to (2n – 1). This series diverges for every finite x, and its meaning as asymptotic expansion is that, for any $N\in\N$ one has

$\mathrm{erfc}(x) = \frac{e^{-x^2}}{x\sqrt{\pi}}\sum_{n=0}^{N-1} (-1)^n \frac{(2n-1)!!}{(2x^2)^n}+ R_N(x) \,$

where the remainder, in Landau notation, is

$R_N(x)=O(x^{-2N+1} e^{-x^2})$ as $x\to\infty$.

Indeed, the exact value of the remainder is

$R_N(x):= \frac{(-1)^N}{\sqrt{\pi}}2^{-2N+1}\frac{(2N)!}{N!}\int_x^\infty t^{-2N}e^{-t^2}\,\mathrm dt,$

which follows easily by induction, writing $e^{-t^2}=-(2t)^{-1}(e^{-t^2})'$ and integrating by parts.

For large enough values of x, only the first few terms of this asymptotic expansion are needed to obtain a good approximation of erfc(x) (while for not too large values of x note that the above Taylor expansion at 0 provides a very fast convergence).

### Continued fraction expansion

A continued fraction expansion of the complementary error function is: [4]

$\mathrm{erfc}(z) = \frac{z}{\sqrt{\pi}}e^{-z^2} \cfrac{a_1}{z^2+ \cfrac{a_2}{1+ \cfrac{a_3}{z^2+ \cfrac{a_4}{1+\dotsb}}}} \qquad a_1 = 1,\quad a_m = \frac{m-1}{2},\quad m \geq 2.$
↑Jump back a section

## Approximation with elementary functions

Abramowitz and Stegun give several approximations of varying accuracy (equations 7.1.25–28). This allows one to choose the fastest approximation suitable for a given application. In order of increasing accuracy, they are:

$\operatorname{erf}(x)\approx 1-\frac{1}{(1+a_1x+a_2x^2+a_3x^3+a_4x^4)^4}$    (maximum error: 5·10−4)

where a1 = 0.278393, a2 = 0.230389, a3 = 0.000972, a4 = 0.078108

$\operatorname{erf}(x)\approx 1-(a_1t+a_2t^2+a_3t^3)e^{-x^2},\quad t=\frac{1}{1+px}$    (maximum error: 2.5·10−5)

where p = 0.47047, a1 = 0.3480242, a2 = −0.0958798, a3 = 0.7478556

$\operatorname{erf}(x)\approx 1-\frac{1}{(1+a_1x+a_2x^2+\cdots+a_6x^6)^{16}}$    (maximum error: 3·10−7)

where a1 = 0.0705230784, a2 = 0.0422820123, a3 = 0.0092705272, a4 = 0.0001520143, a5 = 0.0002765672, a6 = 0.0000430638

$\operatorname{erf}(x)\approx 1-(a_1t+a_2t^2+\cdots+a_5t^5)e^{-x^2},\quad t=\frac{1}{1+px}$    (maximum error: 1.5·10−7)

where p = 0.3275911, a1 = 0.254829592, a2 = −0.284496736, a3 = 1.421413741, a4 = −1.453152027, a5 = 1.061405429

All of these approximations are valid for x ≥ 0. To use these approximations for negative x, use the fact that erf(x) is an odd function, so erf(x) = −erf(−x).

Another approximation is given by

$\operatorname{erf}(x)\approx \sgn(x) \sqrt{1-\exp\left(-x^2\frac{4/\pi+ax^2}{1+ax^2}\right)}$

where

$a = \frac{8(\pi-3)}{3\pi(4-\pi)} \approx 0.140012.$

This is designed to be very accurate in a neighborhood of 0 and a neighborhood of infinity, and the error is less than 0.00035 for all x. Using the alternate value a ≈ 0.147 reduces the maximum error to about 0.00012.[5]

This approximation can also be inverted to calculate the inverse error function:

$\operatorname{erf}^{-1}(x)\approx \sgn(x) \sqrt{\sqrt{\left(\frac{2}{\pi a}+\frac{\ln(1-x^2)}{2}\right)^2 - \frac{\ln(1-x^2)}{a}} -\left(\frac{2}{\pi a}+\frac{\ln(1-x^2)}{2}\right)}.$
↑Jump back a section

## Numerical approximation

Over the complete range of values, there is an approximation with a maximal error of $1.2\cdot10^{-7}$, as follows:[6]

$\operatorname{erf}(x)=\begin{cases} 1-\tau & \mathrm{for\;}x\ge 0\\ \tau-1 & \mathrm{for\;}x < 0 \end{cases}$

with

$\begin{array}{rcl} \tau & = & t\cdot\exp\left(-x^{2}-1.26551223+1.00002368\cdot t+0.37409196\cdot t^{2}+0.09678418\cdot t^{3}\right.\\ & & \qquad-0.18628806\cdot t^{4}+0.27886807\cdot t^{5}-1.13520398\cdot t^{6}+1.48851587\cdot t^7\\ & & \qquad\left.-0.82215223\cdot t^{8}+0.17087277\cdot t^{9}\right) \end{array}$

and

$t=\frac{1}{1+0.5\,|x|}$
↑Jump back a section

## Applications

When the results of a series of measurements are described by a normal distribution with standard deviation $\scriptstyle\sigma$ and expected value 0, then $\scriptstyle\operatorname{erf}\,\left(\,\frac{a}{\sigma \sqrt{2}}\,\right)$ is the probability that the error of a single measurement lies between −a and +a, for positive a. This is useful, for example, in determining the bit error rate of a digital communication system.

The error and complementary error functions occur, for example, in solutions of the heat equation when boundary conditions are given by the Heaviside step function.

↑Jump back a section

## Related functions

The error function is essentially identical to the standard normal cumulative distribution function, denoted Φ, also named norm(x) by software languages, as they differ only by scaling and translation. Indeed,

$\Phi(x) =\frac{1}{\sqrt{2\pi}}\int_{-\infty}^x e^\tfrac{-t^2}{2}\,\mathrm dt = \frac{1}{2}\left[1+\operatorname{erf}\left(\frac{x}{\sqrt{2}}\right)\right]=\frac{1}{2}\,\operatorname{erfc}\left(-\frac{x}{\sqrt{2}}\right)$

or rearranged for erf and erfc:

\begin{align} \mathrm{erf}(x) &= 2 \Phi \left ( x \sqrt{2} \right ) - 1 \\ \mathrm{erfc}(x) &= 2 \Phi \left ( - x \sqrt{2} \right )=2\left(1-\Phi \left ( x \sqrt{2} \right)\right). \end{align}

Consequently, the error function is also closely related to the Q-function, which is the tail probability of the standard normal distribution. The Q-function can be expressed in terms of the error function as

$Q(x) =\frac{1}{2} - \frac{1}{2} \operatorname{erf} \left( \frac{x}{\sqrt{2}} \right)=\frac{1}{2}\operatorname{erfc}\left(\frac{x}{\sqrt{2}}\right).$

The inverse of $\scriptstyle\Phi\,$ is known as the normal quantile function, or probit function and may be expressed in terms of the inverse error function as

$\operatorname{probit}(p) = \Phi^{-1}(p) = \sqrt{2}\,\operatorname{erf}^{-1}(2p-1) = -\sqrt{2}\,\operatorname{erfc}^{-1}(2p).$

The standard normal cdf is used more often in probability and statistics, and the error function is used more often in other branches of mathematics.

The error function is a special case of the Mittag-Leffler function, and can also be expressed as a confluent hypergeometric function (Kummer's function):

$\mathrm{erf}(x)= \frac{2x}{\sqrt{\pi}}\,_1F_1\left(\tfrac12,\tfrac32,-x^2\right).$

It has a simple expression in terms of the Fresnel integral.[further explanation needed]

In terms of the Regularized Gamma function P and the incomplete gamma function,

$\operatorname{erf}(x)=\operatorname{sgn}(x) P\left(\tfrac12, x^2\right)={\operatorname{sgn}(x) \over \sqrt{\pi}}\gamma\left(\tfrac12, x^2\right).$

$\scriptstyle\operatorname{sgn}(x) \$ is the sign function.

### Generalized error functions

Graph of generalised error functions En(x):
grey curve: E1(x) = (1 − e −x)/$\scriptstyle\sqrt{\pi}$
red curve: E2(x) = erf(x)
green curve: E3(x)
blue curve: E4(x)
gold curve: E5(x).

Some authors discuss the more general functions:[citation needed]

$E_n(x) = \frac{n!}{\sqrt{\pi}} \int_0^x e^{-t^n}\,\mathrm dt =\frac{n!}{\sqrt{\pi}}\sum_{p=0}^\infin(-1)^p\frac{x^{np+1}}{(np+1)p!}\,.$

Notable cases are:

• E0(x) is a straight line through the origin: $\scriptstyle E_0(x)=\frac{x}{e \sqrt{\pi}}$
• E2(x) is the error function, erf(x).

After division by n!, all the En for odd n look similar (but not identical) to each other. Similarly, the En for even n look similar (but not identical) to each other after a simple division by n!. All generalised error functions for n > 0 look similar on the positive x side of the graph.

These generalised functions can equivalently be expressed for x > 0 using the Gamma function and incomplete Gamma function:

$E_n(x) = \frac{\Gamma(n)\left(\Gamma\left(\frac{1}{n}\right)-\Gamma\left(\frac{1}{n},x^n\right)\right)}{\sqrt\pi}, \quad \quad x>0.\$

Therefore, we can define the error function in terms of the incomplete Gamma function:

$\operatorname{erf}(x) = 1 - \frac{\Gamma\left(\frac{1}{2},x^2\right)}{\sqrt\pi}.\$

### Iterated integrals of the complementary error function

The iterated integrals of the complementary error function are defined by

$\mathrm i^n \operatorname{erfc}\, (z) = \int_z^\infty \mathrm i^{n-1} \operatorname{erfc}\, (\zeta)\;\mathrm d \zeta.\,$

They have the power series

$\mathrm i^n \operatorname{erfc}\, (z) = \sum_{j=0}^\infty \frac{(-z)^j}{2^{n-j}j! \Gamma \left( 1 + \frac{n-j}{2}\right)}\,,$

from which follow the symmetry properties

$\mathrm i^{2m} \operatorname{erfc} (-z) = - \mathrm i^{2m} \operatorname{erfc}\, (z) + \sum_{q=0}^m \frac{z^{2q}}{2^{2(m-q)-1}(2q)! (m-q)!}$

and

$\mathrm i^{2m+1} \operatorname{erfc} (-z) = \mathrm i^{2m+1} \operatorname{erfc}\, (z) + \sum_{q=0}^m \frac{z^{2q+1}}{2^{2(m-q)-1}(2q+1)! (m-q)!}\,.$
↑Jump back a section

## Implementations

• C: C99 provides the functions double erf(double x) and double erfc(double x) in the header math.h. The pairs of functions {erff(),erfcf()} and {erfl(),erfcl()} take and return values of type float and long double respectively. For complex double arguments, the function names cerf and cerfc are "reserved for future use"; the missing implementation is provided by the open-source project libcerf, which is based on the Faddeeva package.
• C++: C++11 provides erf() and erfc() in the header cmath. Both functions are overloaded to accept arguments of type float, double, and long double. For complex<double>, the above-mentioned C implementations can be used.
• Fortran: The Fortran 2008 standard provides the ERF, ERFC and ERFC_SCALED functions to calculate the error function and its complement for real arguments. Fortran 77 implementations are available in SLATEC.
• Perl: erf (for real arguments, using Cody's algorithm[7]) is implemented in the Perl module Math::SpecFun
• Python: Included since version 2.7 as math.erf() for real arguments. For previous versions or for complex arguments, an implementation of erf for complex arguments is in SciPy as scipy.special.erf().[8] A complex-argument version is also in the arbitrary-precision arithmetic mpmath library as mpmath.erf()
• Mathematica: erf is implemented as Erf and Erfc in Mathematica for real and complex arguments, which are also available in Wolfram Alpha.
• Haskell (programming language): An erf package[9] exists that provides a typeclass for the error function and implementations for the native (real) floating point types.
• R: "The so-called 'error function'"[10] is not provided directly, but is detailed as an example of the normal cumulative distribution function (?pnorm), which is based on W. J. Cody's rational Chebyshev approximation algorithm.[7]
• MathCAD provides both erf(x) and erfc(x) for real arguments.
• Matlab provides both erf and erfc for real arguments, also via W. J. Cody's algorithm.[7]
• Maxima provides both erf and erfc for real and complex arguments.
• Ruby: Provides Math.erf() and Math.erfc() for real arguments.
• Java: Apache commons-math[11] provides implementations of erf and erfc for real arguments.
• Google search: Google's search also acts as a calculator and will evaluate "erf(...)" and "erfc(...)" for real arguments.
↑Jump back a section

## Table of values

x erf(x) erfc(x) x erf(x) erfc(x) 0.00 0.0000000 1.0000000 1.30 0.9340079 0.0659921 0.05 0.0563720 0.9436280 1.40 0.9522851 0.0477149 0.10 0.1124629 0.8875371 1.50 0.9661051 0.0338949 0.15 0.1679960 0.8320040 1.60 0.9763484 0.0236516 0.20 0.2227026 0.7772974 1.70 0.9837905 0.0162095 0.25 0.2763264 0.7236736 1.80 0.9890905 0.0109095 0.30 0.3286268 0.6713732 1.90 0.9927904 0.0072096 0.35 0.3793821 0.6206179 2.00 0.9953223 0.0046777 0.40 0.4283924 0.5716076 2.10 0.9970205 0.0029795 0.45 0.4754817 0.5245183 2.20 0.9981372 0.0018628 0.50 0.5204999 0.4795001 2.30 0.9988568 0.0011432 0.55 0.5633234 0.4366766 2.40 0.9993115 0.0006885 0.60 0.6038561 0.3961439 2.50 0.9995930 0.0004070 0.65 0.6420293 0.3579707 2.60 0.9997640 0.0002360 0.70 0.6778012 0.3221988 2.70 0.9998657 0.0001343 0.75 0.7111556 0.2888444 2.80 0.9999250 0.0000750 0.80 0.7421010 0.2578990 2.90 0.9999589 0.0000411 0.85 0.7706681 0.2293319 3.00 0.9999779 0.0000221 0.90 0.7969082 0.2030918 3.10 0.9999884 0.0000116 0.95 0.8208908 0.1791092 3.20 0.9999940 0.0000060 1.00 0.8427008 0.1572992 3.30 0.9999969 0.0000031 1.10 0.8802051 0.1197949 3.40 0.9999985 0.0000015 1.20 0.9103140 0.0896860 3.50 0.9999993 0.0000007
x erfc(x)/2
1 7.86496e−2
2 2.33887e−3
3 1.10452e−5
4 7.70863e−9
5 7.6873e−13
6 1.07599e−17
7 2.09191e−23
8 5.61215e−30
9 2.06852e−37
10 1.04424e−45
11 7.20433e−55
12 6.78131e−65
13 8.69779e−76
14 1.51861e−87
15 3.6065e−100
16 1.16424e−113
17 5.10614e−128
18 3.04118e−143
19 2.45886e−159
20 2.69793e−176
21 4.01623e−194
22 8.10953e−213
23 2.22063e−232
24 8.24491e−253
25 4.15009e−274
26 2.8316e−296
27 2.61855e−319
↑Jump back a section

### Related functions

#### In probability

↑Jump back a section

## References

1. ^ Andrews, Larry C.; Special functions of mathematics for engineers
2. ^ a b Greene, William H.; Econometric Analysis (fifth edition), Prentice-Hall, 1993, p. 926, fn. 11
3. ^ Van Zeghbroeck, Bart; Principles of Semiconductor Devices, University of Colorado, 2011. [1]
4. ^ Cuyt, Annie A. M.; Petersen, Vigdis B.; Verdonk, Brigitte; Waadeland, Haakon; Jones, William B. (2008). Handbook of Continued Fractions for Special Functions. Springer-Verlag. ISBN 978-1-4020-6948-2.
5. ^ Winitzki, Sergei (6 February 2008). "A handy approximation for the error function and its inverse" (PDF). Retrieved 2011-10-03.
6. ^ Numerical Recipes in Fortran 77: The Art of Scientific Computing (ISBN 0-521-43064-X), 1992, page 214, Cambridge University Press.
7. ^ a b c Cody, William J. (1969). "Rational Chebyshev Approximations for the Error Function". Math. Comp. 23 (107): 631–637. doi:10.1090/S0025-5718-1969-0247736-4.
8. ^ http://docs.scipy.org/doc/scipy/reference/generated/scipy.special.erf.html