The benefit of this approximation is that is converted from an exponent to a multiplicative factor. This can greatly simplify mathematical expressions (as in the example below) and is a common tool in physics.
The approximation can be proven several ways, and is closely related to the binomial theorem. By Bernoulli's inequality, the left-hand side of the approximation is greater than or equal to the right-hand side whenever and .
where and may be real or complex can be expressed as a Taylor Series about the point zero.
If and ≪ , then the terms in the series become progressively smaller and it can be truncated to
.
This result from the binomial approximation can always be improved by
keeping additional terms from the Taylor Series above. This is
especially important when starts to approach one, or when evaluating a more complex expression where the first two terms in the Taylor Series cancel (see example).
Sometimes it is wrongly claimed that ≪ is a sufficient condition for the binomial approximation. A simple counterexample is to let and . In this case but the binomial approximation yields . For small but large , a better approximation is:
Examples
Example simplification
Consider the following expression where and are real but ≫ .
The mathematical form for the binomial approximation can be recovered by factoring out the large term and recalling that a square root is the same as a power of one half.
Evidently the expression is linear in when ≫ which is otherwise not obvious from the original expression.
Example keeping the quadratic term
Consider the expression:
where and ≪ . If only the linear term from the binomial approximation is kept then the expression unhelpfully simplifies to zero
.
While the expression is small, it is not exactly zero. It is
possible to extract a nonzero approximate solution by keeping the
quadratic term in the Taylor Series, i.e. so now,
This result is quadratic in which is why it did not appear when only the linear in terms in were kept.
where an represents the coefficient of the nth term and c is a constant. an is independent of x and may be expressed as a function of n (e.g., ). Power series are useful in analysis since they arise as Taylor series of infinitely differentiable functions. In fact, Borel's theorem implies that every power series is the Taylor series of some smooth function.
In many situations c (the center of the series) is equal to zero, for instance when considering a Maclaurin series. In such cases, the power series takes the simpler form
Any polynomial can be easily expressed as a power series around any center c,
although all but finitely many of the coefficients will be zero since a
power series has infinitely many terms by definition. For instance,
the polynomial can be written as a power series around the center as
or around the center as
or indeed around any other center c. One can view power series as being like "polynomials of infinite degree," although power series are not polynomials.
which is valid for , is one of the most important examples of a power series, as are the exponential function formula
and the sine formula
valid for all real x.
These power series are also examples of Taylor series.
On the set of exponents
Negative powers are not permitted in a power series; for instance, is not considered a power series (although it is a Laurent series). Similarly, fractional powers such as are not permitted (but see Puiseux series). The coefficients are not allowed to depend on , thus for instance:
is not a power series.
Radius of convergence
A power series will converge for some values of the variable x and may diverge for others. All power series f(x) in powers of (x − c) will converge at x = c. (The correct value f(c) = a0 requires interpreting the expression 00 as equal to 1.) If c is not the only convergent point, then there is always a number r with 0 < r ≤ ∞ such that the series converges whenever |x − c| < r and diverges whenever |x − c| > r. The number r is called the radius of convergence of the power series; in general it is given as
For |x − c| = r, we cannot make any general statement on whether the series converges or diverges. However, for the case of real variables, Abel's theorem states that the sum of the series is continuous at x if the series converges at x. In the case of complex variables, we can only claim continuity along the line segment starting at c and ending at x.
Operations on power series
Addition and subtraction
When two functions f and g are decomposed into power series around the same center c, the power series of the sum or difference of the functions can be obtained by termwise addition and subtraction. That is, if
and
then
It is not true that if two power series and have the same radius of convergence, then also has this radius of convergence. If and , then both series have the same radius of convergence of 1, but the series has a radius of convergence of 3.
Multiplication and division
With the same definitions for and , the power series of the product and quotient of the functions can be obtained as follows:
The sequence is known as the convolution of the sequences and .
For division, if one defines the sequence by
then
and one can solve recursively for the terms by comparing coefficients.
Solving the corresponding equations yields the formulae based on determinants of certain matrices of the coefficients of and
Differentiation and integration
Once a function is given as a power series as above, it is differentiable on the interior of the domain of convergence. It can be differentiated and integrated quite easily, by treating every term separately:
Both of these series have the same radius of convergence as the original one.
Analytic functions
A function f defined on some open subsetU of R or C is called analytic if it is locally given by a convergent power series. This means that every a ∈ U has an open neighborhoodV ⊆ U, such that there exists a power series with center a that converges to f(x) for every x ∈ V.
Every power series with a positive radius of convergence is analytic on the interior of its region of convergence. All holomorphic functions
are complex-analytic. Sums and products of analytic functions are
analytic, as are quotients as long as the denominator is non-zero.
If a function is analytic, then it is infinitely differentiable,
but in the real case the converse is not generally true. For an analytic
function, the coefficients an can be computed as
where denotes the nth derivative of f at c, and . This means that every analytic function is locally represented by its Taylor series.
The global form of an analytic function is completely determined by its local behavior in the following sense: if f and g are two analytic functions defined on the same connected open set U, and if there exists an element c∈U such that f (n)(c) = g (n)(c) for all n ≥ 0, then f(x) = g(x) for all x ∈ U.
If a power series with radius of convergence r is given, one can consider analytic continuations of the series, i.e. analytic functions f which are defined on larger sets than { x : |x − c| < r } and agree with the given power series on this set. The number r is maximal in the following sense: there always exists a complex numberx with |x − c| = r such that no analytic continuation of the series can be defined at x.
The
sum of a power series with a positive radius of convergence is an
analytic function at every point in the interior of the disc of
convergence. However, different behavior can occur at points on the
boundary of that disc. For example:
Divergence while the sum extends to an analytic function: has radius of convergence equal to and diverges at every point of . Nevertheless, the sum in is , which is analytic at every point of the plane except for .
Convergent at some points divergent at others.: has radius of convergence . It converges for , while it diverges for
Absolute convergence at every point of the boundary: has radius of convergence , while it converges absolutely, and uniformly, at every point of due to Weierstrass M-test applied with the hyper-harmonic convergent series.
Convergent on the closure of the disc of convergence but not continuous sum: Sierpiński gave an example of a power series with radius of convergence , convergent at all points with ,
but the sum is an unbounded function and, in particular, discontinuous.
A sufficient condition for one-sided continuity at a boundary point is
given by Abel's theorem.
Formal power series
In abstract algebra, one attempts to capture the essence of power series without being restricted to the fields of real and complex numbers, and without the need to talk about convergence. This leads to the concept of formal power series, a concept of great utility in algebraic combinatorics.
Power series in several variables
An extension of the theory is necessary for the purposes of multivariable calculus. A power series is here defined to be an infinite series of the form
where j = (j1, ..., jn) is a vector of natural numbers, the coefficients a(j1, …, jn) are usually real or complex numbers, and the center c = (c1, ..., cn) and argument x = (x1, ..., xn) are usually real or complex vectors. The symbol is the product symbol, denoting multiplication. In the more convenient multi-index notation this can be written
where is the set of natural numbers, and so is the set of ordered n-tuples of natural numbers.
The theory of such series is trickier than for single-variable
series, with more complicated regions of convergence. For instance, the
power series is absolutely convergent in the set between two hyperbolas. (This is an example of a log-convex set, in the sense that the set of points , where
lies in the above region, is a convex set. More generally, one can
show that when c=0, the interior of the region of absolute convergence
is always a log-convex set in this sense.) On the other hand, in the
interior of this region of convergence one may differentiate and
integrate under the series sign, just as one may with ordinary power
series.
Order of a power series
Let α be a multi-index for a power series f(x1, x2, ..., xn). The order of the power series f is defined to be the least value such that there is aα ≠ 0 with , or if f ≡ 0. In particular, for a power series f(x) in a single variable x, the order of f is the smallest power of x with a nonzero coefficient. This definition readily extends to Laurent series.