In line with the general purpose of this thread, in this post I intend to present a bird’s eye view of the concept of power expansions (Taylor series), in the belief that these ideas are often presented in a highly technical way that makes it difficult for students to grasp their simplicity and significance.
The change of the area of a unit square corresponding to a change of its side equals
The linear part of the change is called the first differential of (to be precise, the first differential of
at the point
, corresponding to a change
of the independent variable). It is denoted by
:
Thus,
.
At this point, the above relation is exact and is valid for any finite . However, if we consider an infinitesimal
, the main contribution to
is given by the linear (in
) part
, the difference being an infinitesimal of higher order. This observation suggests the following definition: if a function
of a variable
is such that
,
for infinitesimal increments , we say that
is differentiable at
. We define the first differential to be
. The coefficient
is what we call the derivative of
at the point
. Therefore, the derivative is a ratio of differentials,
. Finding the derivative from the knowledge of
involves, in the general case, considering
and
infinitesimal (see below). For polynomials, however, it only requires algebraic manipulations.
For example, let us compute the derivative of at a generic
. We have to increase
by
and compute the corresponding change of
:
The linear part is
while the rest are quadratic and cubic terms, all of them infinitesimals of higher order, negligible w.r.t. . Thus, the derivative of
with respect to
is
In both previous examples and
, the total variations of
(respectively
) were sums of positive integer powers of
. That is always the case when considering polynomial functions of
. Namely, if
is a polynomial of degree
, and we look at the change of
as
is increased from
to
, we will have
with coefficients depending on the base point
. In order to get
, we just expand the powers
according to Newton’s binomial theorem. We can think of the coefficients as rates of different orders, multiplying the corresponding powers of
. Thus, we have a clear separation of the factors leading to the change
: the rates, which depend only on the base point, and the different powers of the variation of the independent variable
. This is a simple instance of a (finite) “Taylor expansion”. No limit process is involved in the previous computations, and the final expression
is exact for any finite value of
.
One of the main breakthroughs in the development of Calculus is the fact that for non-polynomial functions like or
a similar expansion holds. The key difference is that the expansion may contain infinitely many terms. Thus, the concept of a power series serves as a bridge between algebraic and transcendental relations. The idea that (at least smooth) functional relations are either polynomials or “infinite degree polynomials” (power series) dominated Analysis over a long period of time. Newton himself considered the binomial series and the use of series in general to solve differential equations his main mathematical achievement. They also allowed to define many non-elementary functions and to develop Complex Analysis. Series were informally used by Euler, Lagrange, Laplace and many others, in some cases producing paradoxical results. Questions related to convergence were not posed until the middle of the XIX century by Gauss, Abel, Cauchy, etc. It is said that when Laplace heard about Cauchy’s convergence criteria, he rushed home to check the series he used in his monumental “Celestial Mechanics”. Luckily for him and for the stability of the Solar System, all of them were convergent in the range of parameters he considered.
For a relation like , algebraic methods to find the coefficients in the expansion are not available, and need to be replaced by limit procedures. Here is the basic idea: first, we divide throughout by
, yielding
where the denote etc. simply as
etc. As
approaches zero, all the terms in the r.h.s. vanish except for
. Consequently,
is the limit value of
. This is how the (first) derivative is usually defined,
We can think of successive terms in the right-hand side of as corrections to the previous ones when
is infinitesimal. Thus, if we only keep the first differential, we obtain the linear approximation of
near
,
which renders a good estimate if is small enough, since we are discarding infinitesimals of higher order.
In order to find the coefficient above, we repeat the procedure. Namely, we take
to the left in (!) and divide by
again, yielding
Thus, the infinitesimal is typically quadratic,
or of a higher order,
if
. It is natural that the quadratic correction
is related to the variation of
or, equivalently,
over the interval
. Thus, we introduce the second differential of
to measure the variation of the first differential as the base point
changes to
. Namely, we define
where we consider a constant and vary the base point
again by the amount
(this is reminiscent of arithmetic sequences). We have
,
and therefore
,
where is the derivative of the first derivative
at
. Thus the second differential is quadratic in
. The coefficient
is called the second derivative of
at the point
The second derivative is thus the ratio of the second differential to the square of
.
How is related to
? The idea is to move along the linear approximation up to
and, at that point, use
to “correct” the linear approximation. From the above expression for
,
,
where the dots represent infinitesimals of order higher than . In the limit
we obtain
,
that is, .
A completely analogous procedure allows to find in
. This time we need to split the interval
into three equal pieces, since we want to account for the variation of the second differential, which depends on three points. Precisely, we introduce the third differential of
at
as:
.
A computation, similar to the one above for the second differential, gives
,
where is called the third derivative of
at
. The expression for
is
A computation similar to the one above gives the relation . In general, we define the
-th differential recursively
and the corresponding -th derivative
. The total variation of a polynomial of degree
can then be expressed as
.
Thus, if we develop a machinery to compute derivatives, we can swiftly write down the expansion for a polynomial, avoiding the use of the binomial theorem and rearrangement. More importantly, transcendental functions like
or
can be dealt with the same way, leading to their power (Taylor) expansions:
For computational purposes, the more terms we keep on the right hand side, the better the approximation for small values of . But series expansions are also important for the theory. Many of the usual manipulations with polynomials can be extended to series, including differentiation, integration, long division, etc. In particular, they can be used to solve differential equations with “nice” coefficients.
However, the transition from polynomials to more general functions is non-trivial. We showed above that if the difference can be expressed in the form
, then
has to be a differentiable function of
and the coefficients
, etc are uniquely determined in terms of the derivatives. The question as to whether such a representation is at all available, even for infinitely differentiable functions, leads to the concept of analytic function, to be considered in a future post.