A slant asymptote for a real function of one real variable is a line
with the property
,
where as
. Geometrically, the graph of
comes closer and closer to the line for large positive/negative values of
. For the sake of simplicity, we will deal with asymptotes at
, the other case being completely analogous. When
we say that the asymptote is horizontal. A simple way to detect if a given function has a slant asymptote is by checking linearity at infinity,
if the limit
is finite. If that is the case, the value of the limit is the slope of the asymptote, . The free term
is then given by the limit
,
if the latter exists (and is finite). For rational functions of the form
where and
are polynomials, the situation is much simpler. In order for
to be linear at infinity, we need
and, if that is the case, we perform long division which leads to
where and, therefore, the asymptote is just the quotient
since
When , the fraction is superlinear at infinity.
All this is well known and usually taught in high school.
On the other hand, students are taught to find tangents to a given curve at a given point by means of derivatives. It turns out, however, that finding the tangent to a rational function at does not require derivatives at all. In order to understand this, we notice that
is a tangent to
at
precisely when
with as
. The last relation is very similar to
except for the fact that now we are looking at
instead of
. Thus, if we could somehow come up with a relation like
where now
,
the quotient would be the tangent.
As it happens, that is perfectly possible. All we need to do is divide the polynomials starting with the lowest powers (backwards), until we reach a “partial remainder” whose lowest degree is two or higher. Observe that the lowest degree of the divisor is necessarily zero (otherwise
is undefined at
).
As an example, let us find the tangent to the graph of
at . Starting with the lowest degree, we have
,
with first partial remainder
. In the next step, we add
to the quotient, with a partial remainder
. Since we are interested in the tangent line and
is an infinitesimal of degree higher than one at zero, the division stops here and the equation of the tangent is
. If we keep dividing, we get the Taylor polynomials of higher degree. For instance, the osculating parabola at
is
and so on.
Long division is taught at school starting with the highest powers. A possible reason is that long division of numbers proceeds by reducing the remainder at each step. If we replace the base in the decimal representation of numbers by
, we arrive at the usual long division algorithms of polynomials, reducing the degree at each step. I would call this procedure “division at infinity”. In contrast, the above is an example of “division at zero”.
Finding the tangent to a rational function at a point different from zero can be reduced to the previous case. If, say, we need to find the tangent to at
, all we need to do is set
and express
and
as polynomials in
and then finding the tangent at
as before. Finally, we have to replace
back by
in the found equation of the tangent.
The above reveals a perfect symmetry between the problems of finding the asymptote and that of finding the tangent at . In some sense, we can say that an asymptote is a “tangent at infinity” and, I guess, that a tangent is an asymptote at a finite point. Both problems are algebraic in nature and can be solved without limit procedures, just by means of division (forward or backward).
More generally, for rational functions, being the simplest non-polynomial functions, finding their Taylor expansion at zero (and, by translation at any point) is a pure algebraic procedure. I believe this fact should be emphasized in high school and it could be used as a motivating example to introduce more general power expansions. As a matter of fact, Newton was inspired by the algorithm of long division for numbers to start experimenting with power series, not necessarily with integer powers. The image below shows a page from his “Method of Fluxions and Infinite Series”. The “backwards” long division of by
is performed in order to get the power series expansion.

The differential of a quotient
Here is yet another example of “division at zero”.
When students are exposed to the differentiation rules, those are derived from the definition of derivative as the limit of the differential quotient. Thus for example to prove the rule of differentiation of a product we proceed as follows.
given that all the limits are assumed to exist. A similar computation can be done for the quotient. It should be noted, however, that a little algebraic trick has to be used in both cases to make the derivatives of the individual factors appear explicitly. No big deal, but a bit artificial. And, importantly, not the way the founders of Infinitesimal Calculus arrived at these rules.
To help intuition, the product rule is often presented in the form
,
and the last term is ignored in the last equality as being a quadratic infinitesimal (in Leibniz’ terminology, the last equality is actually an “adequality”, a term coined by Fermat). Without a doubt, the latter derivation, albeit not meeting the modern standards of rigor, reveals the reason for the presence of the “mixed” terms and the general structure of the formula. Moreover, no algebraic tricks are required. The formula follows in a straightforward manner.
A similar derivation of the quotient rule involves “division at zero”. Here is the derivation in the book “Calculus made easy” by Silvanus Thompson, from 1910.

Observe that the operation has been stopped when the remainder is a quadratic infinitesimal. The conclusion of the computation is the familiar rule