English | Español

Try our Free Online Math Solver!

Online Math Solver

 

 

 

 

 

 

 

 
 
 
 
 
 
 
 
 

 

 

 
 
 
 
 
 
 
 
 

Please use this form if you would like
to have this math solver on your website,
free of charge.


Exponential and Log Functions

Logarithms
First we discuss the inverse function of an exponential function , the so called logarithm
function. It follows from arguments like those above that an exponential function ax with a >0, is
increasing if and only if a > 1, and is decreasing if a < 1. The exponential function 1x is neither
increasing nor decreasing, but a constant equal to 1, and has no inverse. However for all a > 0 and
a ≠ 1, the function ax has an inverse function called the log base a, written . To find the
domain of the log function we must determine the values of the exponential function, so we
assume for ease that a > 1. Then notice that if say a = 1+h where h > 0, then for all positive
integers n, we have with all terms positive , so an > 1+nh. Since the
right hand side grows to infinity as n does, we conclude that ax gets arbitrarily large for large x,
i.e. the limit of ax is +∞ as x approaches +∞. Since , we see that ax approaches zero
from above as x approaches -∞. Thus ax assumes all positive values as x ranges over all reals, so
the domain of the inverse function is the positive reals . Moreover is also continuous
and satisfies the law for all positive reals x , y, opposite to the law for
the exponential
function. Since a function determines its inverse, we have the analogous theorem
to recognize a log function.

Theorem: If g is a non constant continuous function defined for all positive reals, satisfying
1) g(1)= 0, and
2) g(xy) = g(x) + g(y), for all positive x , y,
then there is a unique a > 0 with g(a) = 1, and , for all positive x.

Now all we have to do is find a differentiable function g satisfying the conditions in the theorem,
and then it must be a log function. if we find one with g(2) = 1, it will prove that is
differentiable. Moreover by the inverse function theorem it will follow that 2x is also
differentiable, and we will have accomplished by a very indirect route, the goal of proving this
which we temporarily gave up on earlier.

The only tool we have for constructing differentiable functions is the fundamental theorem of
calculus, which allows us to construct a function with any given continuous derivative. Thus in
order to construct a log function we need to know the derivative of a log function. Using the
chain rule we have so so Thus
assuming they are differentiable, a log function must have derivative equal to 1/Kx. Now with
this information, we can construct a differentiable function with derivative 1/Kx, and ask
whether it is a log function, which we have every right to expect to be true. Moreover the
simplest choice for k is obviously 1, so we begin from that choice.

Define . We claim L(x) is a log function. To check this we must show that L(1) = 0
(which is obvious), then that L is continuous, which is always true for a function defined by an
integral, indeed by the FTC this one is even differentiable with derivative 1/x, and finally that
L(xy) = L(x) + L(y) for all x,y > 0. To prove this last formula we use the MVT . I.e. fix y > 0
and let g(x) = L(xy). Then g'(x) = yL'(xy) = y(1/xy) = 1/x = L'(x). Thus L and g have the same
derivative so must differ by a constant according to the MVT. I.e. g(x) = L(xy) = L(x) + c for
some c. To evaluate c, set x = 1 and get L(y) = L(1) + c = c. So c = L(y) and thus L(xy) = L(x) +
L(y) as claimed.

To see what the base is we must find the unique positive number a such that L(a) = 1.
This is not so immediate, but one can show using approximations of the integral that the base is a
number we shall call e such that 2.71828 < e < 2.71829. Indeed since the midpoint estimate is an
underestimate for a function like 1/x which is concave up, using the midpoint estimate on the
interval [1,3] we get L(3) ≥ 2(1/2) = 1, so e ≤ 3. Subdividing into [1,2] and [2,2.8] we have
midpoint estimate 2/3 + 5/12 = 13/12 ≤ L(2.8), so e < 2.8. Then using the trapezoidal upper
estimate for the subdivision [1,1.4], [1.4,1.8] and [1.8,2.2], [2.2,2.6] gives (1/5)(1 + 10/7 + 10/9
+10/11 + 5/13) ≤ (1/5) (1 + 1.43 + 1.112 + .9091 + .385) ≤ .97. Thus e > 2.6. Later we will find
a better way to estimate e using infinite series.

We see that but it is usual to write it as ln(x). In particular ln(x) is a
differentiable function defined on all positive reals, with derivative 1/x, and which takes on all real
values. Its inverse is the exponential function ex whose derivative is also ex. If K = ln(2) and if
we consider the function , it follows from the same argument as above using MVT,
that g is a logarithm function. The base is the unique number a such that g(a) = 1. But since g(x)
= (1/K)ln(x) = ln(x)/ln(2), it follows that this number is 2. Hence , and we see the
derivative of g is 1/(ln(2)x). Using the inverse function theorem, the inverse of g is the
differentiable function f(x) = 2x with derivative f'(x) = ln(2) 2x.
We obtain as well that for all a > 0, the function where K = ln(a), is the
differentiable log function . Its inverse is then the differentiable function ax, with
derivative ln(a) ax. In fact, for all a > 0, we can express ax in terms of ex, by using the corollary
above to write

Now that we have definition and the basic properties of exponential and log functions, the other
basic properties are their rates of growth. According to the formulas above we can express every
exponential function in terms of the simplest one ex, so we may consider only that one. Similarly
we saw above that for all a > 0, that Basically ex grows rapidly, faster than
any positive power of x , and ln(x) grows slowly, slower than any positive power of x.
More precisely , for all r > 0, we have , and .
There are several ways to prove this, but the easiest way, and the way which teaches us the most
useful fact, is to use l'Hopital's rule.

Theorem: If f, g are both differentiable functions such that both have limit ± ∞, or if both have
limit 0, as x approaches a, or as x approaches ∞, then the limit of f/g equals the limit of f'/g',
assuming that latter limit exists.

Using this theorem on , we can take derivatives until the power of x in the top is non
positive, so that then the function in the top is constant or approaches zero as x approaches
infinity, and the one in the bottom, still equal to ex, approaches infinity. Then the limit is 1/∞, or
0/∞, both = 0. For the limit , one can begin with ln(x)/x and take one derivative in top and
bottom to finish. Then for the general case, one may consider instead the equivalent limit
which we have already done. (The only point is that the bottom
number should be the rth power of the top number, and that both approach infinity, it does not
matter which one we call x.)

Prev Next