# Convergent series

(Redirected from Convergence (mathematics))

In mathematics, a series is the sum of the terms of an infinite sequence of numbers. More precisely, an infinite sequence ${\displaystyle (a_{0},a_{1},a_{2},\ldots )}$ defines a series S that is denoted

${\displaystyle S=a_{0}+a_{1}+a_{2}+\cdots =\sum _{k=0}^{\infty }a_{k}.}$

The nth partial sum Sn is the sum of the first n terms of the sequence; that is,

${\displaystyle S_{n}=\sum _{k=1}^{n}a_{k}.}$

A series is convergent (or converges) if the sequence ${\displaystyle (S_{1},S_{2},S_{3},\dots )}$ of its partial sums tends to a limit; that means that, when adding one ${\displaystyle a_{k}}$ after the other in the order given by the indices, one gets partial sums that become closer and closer to a given number. More precisely, a series converges, if there exists a number ${\displaystyle \ell }$ such that for every arbitrarily small positive number ${\displaystyle \varepsilon }$, there is a (sufficiently large) integer ${\displaystyle N}$ such that for all ${\displaystyle n\geq N}$,

${\displaystyle \left|S_{n}-\ell \right|<\varepsilon .}$

If the series is convergent, the (necessarily unique) number ${\displaystyle \ell }$ is called the sum of the series.

The same notation

${\displaystyle \sum _{k=1}^{\infty }a_{k}}$

is used for the series, and, if it is convergent, to its sum. This convention is similar to that which is used for addition: a + b denotes the operation of adding a and b as well as the result of this addition, which is called the sum of a and b.

Any series that is not convergent is said to be divergent or to diverge.

## Examples of convergent and divergent series

• The reciprocals of the positive integers produce a divergent series (harmonic series):
${\displaystyle {1 \over 1}+{1 \over 2}+{1 \over 3}+{1 \over 4}+{1 \over 5}+{1 \over 6}+\cdots \rightarrow \infty .}$
• Alternating the signs of the reciprocals of positive integers produces a convergent series (alternating harmonic series):
${\displaystyle {1 \over 1}-{1 \over 2}+{1 \over 3}-{1 \over 4}+{1 \over 5}-\cdots =\ln(2)}$
• The reciprocals of prime numbers produce a divergent series (so the set of primes is "large"; see divergence of the sum of the reciprocals of the primes):
${\displaystyle {1 \over 2}+{1 \over 3}+{1 \over 5}+{1 \over 7}+{1 \over 11}+{1 \over 13}+\cdots \rightarrow \infty .}$
• The reciprocals of triangular numbers produce a convergent series:
${\displaystyle {1 \over 1}+{1 \over 3}+{1 \over 6}+{1 \over 10}+{1 \over 15}+{1 \over 21}+\cdots =2.}$
• The reciprocals of factorials produce a convergent series (see e):
${\displaystyle {\frac {1}{1}}+{\frac {1}{1}}+{\frac {1}{2}}+{\frac {1}{6}}+{\frac {1}{24}}+{\frac {1}{120}}+\cdots =e.}$
• The reciprocals of square numbers produce a convergent series (the Basel problem):
${\displaystyle {1 \over 1}+{1 \over 4}+{1 \over 9}+{1 \over 16}+{1 \over 25}+{1 \over 36}+\cdots ={\pi ^{2} \over 6}.}$
• The reciprocals of powers of 2 produce a convergent series (so the set of powers of 2 is "small"):
${\displaystyle {1 \over 1}+{1 \over 2}+{1 \over 4}+{1 \over 8}+{1 \over 16}+{1 \over 32}+\cdots =2.}$
• The reciprocals of powers of any n>1 produce a convergent series:
${\displaystyle {1 \over 1}+{1 \over n}+{1 \over n^{2}}+{1 \over n^{3}}+{1 \over n^{4}}+{1 \over n^{5}}+\cdots ={n \over n-1}.}$
• Alternating the signs of reciprocals of powers of 2 also produces a convergent series:
${\displaystyle {1 \over 1}-{1 \over 2}+{1 \over 4}-{1 \over 8}+{1 \over 16}-{1 \over 32}+\cdots ={2 \over 3}.}$
• Alternating the signs of reciprocals of powers of any n>1 produces a convergent series:
${\displaystyle {1 \over 1}-{1 \over n}+{1 \over n^{2}}-{1 \over n^{3}}+{1 \over n^{4}}-{1 \over n^{5}}+\cdots ={n \over n+1}.}$
• The reciprocals of Fibonacci numbers produce a convergent series (see ψ):
${\displaystyle {\frac {1}{1}}+{\frac {1}{1}}+{\frac {1}{2}}+{\frac {1}{3}}+{\frac {1}{5}}+{\frac {1}{8}}+\cdots =\psi .}$

## Convergence tests

There are a number of methods of determining whether a series converges or diverges.

If the blue series, ${\displaystyle \Sigma b_{n}}$ , can be proven to converge, then the smaller series, ${\displaystyle \Sigma a_{n}}$  must converge. By contraposition, if the red series ${\displaystyle \Sigma a_{n}}$  is proven to diverge, then ${\displaystyle \Sigma b_{n}}$  must also diverge.

Comparison test. The terms of the sequence ${\displaystyle \left\{a_{n}\right\}}$  are compared to those of another sequence ${\displaystyle \left\{b_{n}\right\}}$ . If,

for all n, ${\displaystyle 0\leq \ a_{n}\leq \ b_{n}}$ , and ${\displaystyle \sum _{n=1}^{\infty }b_{n}}$  converges, then so does ${\displaystyle \sum _{n=1}^{\infty }a_{n}.}$

However, if,

for all n, ${\displaystyle 0\leq \ b_{n}\leq \ a_{n}}$ , and ${\displaystyle \sum _{n=1}^{\infty }b_{n}}$  diverges, then so does ${\displaystyle \sum _{n=1}^{\infty }a_{n}.}$

Ratio test. Assume that for all n, ${\displaystyle a_{n}}$  is not zero. Suppose that there exists ${\displaystyle r}$  such that

${\displaystyle \lim _{n\to \infty }\left|{\frac {a_{n+1}}{a_{n}}}\right|=r.}$

If r < 1, then the series is absolutely convergent. If r > 1, then the series diverges. If r = 1, the ratio test is inconclusive, and the series may converge or diverge.

Root test or nth root test. Suppose that the terms of the sequence in question are non-negative. Define r as follows:

${\displaystyle r=\limsup _{n\rightarrow \infty }{\sqrt[{n}]{|a_{n}|}},}$
where "lim sup" denotes the limit superior (possibly ∞; if the limit exists it is the same value).

If r < 1, then the series converges. If r > 1, then the series diverges. If r = 1, the root test is inconclusive, and the series may converge or diverge.

The ratio test and the root test are both based on comparison with a geometric series, and as such they work in similar situations. In fact, if the ratio test works (meaning that the limit exists and is not equal to 1) then so does the root test; the converse, however, is not true. The root test is therefore more generally applicable, but as a practical matter the limit is often difficult to compute for commonly seen types of series.

Integral test. The series can be compared to an integral to establish convergence or divergence. Let ${\displaystyle f(n)=a_{n}}$  be a positive and monotonically decreasing function. If

${\displaystyle \int _{1}^{\infty }f(x)\,dx=\lim _{t\to \infty }\int _{1}^{t}f(x)\,dx<\infty ,}$

then the series converges. But if the integral diverges, then the series does so as well.

Limit comparison test. If ${\displaystyle \left\{a_{n}\right\},\left\{b_{n}\right\}>0}$ , and the limit ${\displaystyle \lim _{n\to \infty }{\frac {a_{n}}{b_{n}}}}$  exists and is not zero, then ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  converges if and only if ${\displaystyle \sum _{n=1}^{\infty }b_{n}}$  converges.

Alternating series test. Also known as the Leibniz criterion, the alternating series test states that for an alternating series of the form ${\displaystyle \sum _{n=1}^{\infty }a_{n}(-1)^{n}}$ , if ${\displaystyle \left\{a_{n}\right\}}$  is monotonically decreasing, and has a limit of 0 at infinity, then the series converges.

Cauchy condensation test. If ${\displaystyle \left\{a_{n}\right\}}$  is a positive monotone decreasing sequence, then ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  converges if and only if ${\displaystyle \sum _{k=1}^{\infty }2^{k}a_{2^{k}}}$  converges.

## Conditional and absolute convergence

For any sequence ${\displaystyle \left\{a_{1},\ a_{2},\ a_{3},\dots \right\}}$ , ${\displaystyle a_{n}\leq \left|a_{n}\right|}$  for all n. Therefore,

${\displaystyle \sum _{n=1}^{\infty }a_{n}\leq \sum _{n=1}^{\infty }\left|a_{n}\right|.}$

This means that if ${\displaystyle \sum _{n=1}^{\infty }\left|a_{n}\right|}$  converges, then ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  also converges (but not vice versa).

If the series ${\displaystyle \sum _{n=1}^{\infty }\left|a_{n}\right|}$  converges, then the series ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  is absolutely convergent. The Maclaurin series of the exponential function is absolutely convergent for every complex value of the variable.

If the series ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  converges but the series ${\displaystyle \sum _{n=1}^{\infty }\left|a_{n}\right|}$  diverges, then the series ${\displaystyle \sum _{n=1}^{\infty }a_{n}}$  is conditionally convergent. The Maclaurin series of the logarithm function ${\displaystyle \ln(1+x)}$  is conditionally convergent for x = 1.

The Riemann series theorem states that if a series converges conditionally, it is possible to rearrange the terms of the series in such a way that the series converges to any value, or even diverges.

## Uniform convergence

Let ${\displaystyle \left\{f_{1},\ f_{2},\ f_{3},\dots \right\}}$  be a sequence of functions. The series ${\displaystyle \sum _{n=1}^{\infty }f_{n}}$  is said to converge uniformly to f if the sequence ${\displaystyle \{s_{n}\}}$  of partial sums defined by

${\displaystyle s_{n}(x)=\sum _{k=1}^{n}f_{k}(x)}$

converges uniformly to f.

There is an analogue of the comparison test for infinite series of functions called the Weierstrass M-test.

## Cauchy convergence criterion

The Cauchy convergence criterion states that a series

${\displaystyle \sum _{n=1}^{\infty }a_{n}}$

converges if and only if the sequence of partial sums is a Cauchy sequence. This means that for every ${\displaystyle \varepsilon >0,}$  there is a positive integer ${\displaystyle N}$  such that for all ${\displaystyle n\geq m\geq N}$  we have

${\displaystyle \left|\sum _{k=m}^{n}a_{k}\right|<\varepsilon ,}$

which is equivalent to

${\displaystyle \lim _{n\to \infty \atop m\to \infty }\sum _{k=n}^{n+m}a_{k}=0.}$