In mathematics, an alternating series is an infinite series of terms that alternate between positive and negative signs. In capital-sigma notation this is expressed or with for all .
Like any series, an alternating series is a convergent series if and only if the sequence of partial sums of the series converges to a limit. The alternating series test guarantees that an alternating series is convergent if the terms converge to 0 monotonically, but this condition is not necessary for convergence.
The geometric series − + − + ⋯ sums to .
\pi | |
4 |
The Mercator series provides an analytic power series expression of the natural logarithm, given by
The functions sine and cosine used in trigonometry and introduced in elementary algebra as the ratio of sides of a right triangle can also be defined as alternating series in calculus. andWhen the alternating factor is removed from these series one obtains the hyperbolic functions sinh and cosh used in calculus and statistics.
For integer or positive index α the Bessel function of the first kind may be defined with the alternating series where is the gamma function.
If is a complex number, the Dirichlet eta function is formed as an alternating seriesthat is used in analytic number theory.
See main article: Alternating series test.
The theorem known as the "Leibniz Test" or the alternating series test states that an alternating series will converge if the terms converge to 0 monotonically.
Proof: Suppose the sequence
an
m
m<n
Sn-Sm\leam
Since
an
-(am-am+1)
Sn-Sm\leam
-am\leSn-Sm
am
0
Sm
m
The estimate above does not depend on
n
an
1-1/2+1/3-1/4+...=ln2
a20000
a10000
an-an+1
A series converges absolutely if the series converges.
Theorem: Absolutely convergent series are convergent.
Proof: Suppose is absolutely convergent. Then, is convergent and it follows that converges as well. Since , the series converges by the comparison test. Therefore, the series converges as the difference of two convergent series .
A series is conditionally convergent if it converges but does not converge absolutely.
For example, the harmonic seriesdiverges, while the alternating versionconverges by the alternating series test.
For any series, we can create a new series by rearranging the order of summation. A series is unconditionally convergent if any rearrangement creates a series with the same convergence as the original series. Absolutely convergent series are unconditionally convergent. But the Riemann series theorem states that conditionally convergent series can be rearranged to create arbitrary convergence.[4] Agnew's theorem describes rearrangements that preserve convergence for all convergent series. The general principle is that addition of infinite sums is only commutative for absolutely convergent series.
For example, one false proof that 1=0 exploits the failure of associativity for infinite sums.
As another example, by Mercator series
But, since the series does not converge absolutely, we can rearrange the terms to obtain a series for :
In practice, the numerical summation of an alternating series may be sped up using any one of a variety of series acceleration techniques. One of the oldest techniques is that of Euler summation, and there are many modern techniques that can offer even more rapid convergence.