The Exponential and Logarithmic Functions
2025-04-04 | 2025-04-04
Estimated Reading Time: 31 minutes
This is another in my series of blogs on fascinating and mathematically indispensable numbers. It follows on from blogs on zero, one, and π, and is likely to be followed by others. It happens that a single blog is sometimes too short to display the beauty of the subject, and I have had to segment the story into parts. Such will be the case here. While e is less well known to the general public than π, it is perhaps even more fundamental to all of Nature and pervades the entire realm of Mathematics. It would indeed be difficult to discover a nook or cranny of Nature that has not been penetrated by this omnipresent emissary of mathematical order.
Unfurling countless digits
Perversely, almost all important numbers like
“What if I were the creator of such a virtual world, populated like ours, by irrational numbers with unending and unpredictable digits? How would I sustain that world without an infinite memory to hold all those countless digits?”
I would need some convenient, succinct, shorthand method by which to unfurl their countless digits, one after the other. It might be an algorithm like a convergent infinite series or a recursive definition or an infinite continued fraction1.
This thought is a preface to many of the fascinating numbers we will encounter in these blogs.
I am opening this blog with an abrupt exposure to the idea of
exponentials, without any courteous introduction or gentle historical
note on
Bases and Exponents
We have introduced the different types of numbers in the blog The Two Most Important Numbers: Zero and One. In that very same blog, we also introduced the idea of exponentiation, or raising (something) to a power, as repeated multiplication. That section is very important: do take a look at it again if it seems faint or foggy now, as some basic results from that blog are worth reviewing at this point.
Monomial power functions
At the very outset, it is important to clear up a possible source of confusion: monomial power functions and exponentials might look similar but are very different.
A monomial power function is a monomial
The following points should be noted:
In each case,
varies, but is constant, as defined in Equation 1. When
is even, like , etc., the graph of is symmetrical about the -axis. Such a function is called an even function, defined as . When
is odd, like etc., the graph of exhibits rotational symmetry about the origin , i.e., if the graph is rotated 180° about the origin, the graph remains unchanged. Such a function is called an odd function, defined as . The graph of
is constant and its behaviour is anomalous when compared to others in the family, as is apparent from Figure 1. The higher the value of
the steeper the graph climbs as increases. Except for
, the graphs of pass through for all other values of . The monomial power functions are a subset of the polynomials.
As an exception, I have included in Figure 1 the special case of the positive non-integer power
, which is the subject of this blog. This was simply to show that since lies between and its graph is sandwiched between the curves and . It is shown as a dashed line in Figure 1. But there ends the similarity. In fact, is not a monomial power function. Negative numbers cannot be raised to non-integer powers and still remain real numbers. So, the domain for alone is restricted to . If you find all this unhelpful or confusing, simply ignore it for now.
Exponentials
We now consider the second family of functions which might look like
the monomial power functions but are really a bird of a different
feather. The exponentials are generally defined as:
The following points are noteworthy:
The graph for
is anomalous and constant in value. It is shown only for completeness and may be excluded from the definition of exponentials as in Equation 2. All other graphs pass through the point
, which is characteristic of all exponentials. For
, the values of are greater than , but less than , and approach the asymptote as . As
increases without bound, so does . The larger
is the steeper the rise of for values of .𝑥 > 1 The graph of
—shown as a dashed line—legitimately belongs to this class of curves and shares the same domain as other exponentials. Even as𝑒 𝑥 = e x p 𝑥 , its graph is sandwiched between those of2 < 𝑒 < 3 and2 𝑥 as would be expected.3 𝑥 The exponentials are neither odd nor even functions.
Note how these exponential functions increase far more rapidly than the monomial power functions. The roles of
and𝑛 have been interchanged between the monomial power functions and the exponentials.𝑥
A tabular comparison of the values of
Computational complexity theory
I am belabouring this distinction between the polynomials (or monomial power functions) and the exponentials because many students, especially of computer science, are usually clueless when they encounter the rather forbidding topic called Computational complexity theory in their university studies.
The exponential functions tend to increase extremely rapidly
compared to the polynomial functions. Such distinctions become vital
when evaluating the efficiency and execution times of algorithms in
computer science, and indeed even their solvability in finite time. Keep
this difference in mind as we navigate our way through the number
Introduction to the number 𝑒
We are now ready to make our formal acquaintance with the number
Unlike
The number
Among the important numbers of mathematics, the linkage between
While
The power of the exponent
Did you read that heading carefully? And did you get the pun in it?
We have already peeked into exponentiation in Table 1. Just as multiplication is a shorthand for repeated addition so too is exponentiation a shorthand for repeated multiplication. It has been said that human beings are not very good when it comes to comprehending the very large and the very small.
If I gave you a stick that is one metre long and told you to divide it into one thousand equal parts, how long would each division be? If I now told you that the same stick represented one million divisions, and asked you to mark the first one thousandth part, where would you mark it?
I am not going to tell you, because this one is easy enough for you to figure out for yourself. It will tell you how good or bad your ability to estimate is. What happens if the scale is not linear but logarithmic? Let your mental cogwheels again start turning. If you find all this too exhausting, simply look at Figure 3 below.
The power of two
There is a famous story about the person who invented the game of chess.2 The monarch of the realm was so pleased with the game that he wanted to reward the inventor. Feeling very expansive, he said “Ask for anything and I will give it to you.” The inventor rather diffidently asked the king for one grain of rice on the first square of the chess board, double that number of grains on the second, double that number of grains again on the third, and so on till all the sixty four squares had their quotas filled [cole-1998].

The king laughed and said, “Ask for something more. You deserve it.” The inventor quietly but persistently said, “Sire, kindly grant me what I have asked.” The king jovially asked his ministers to fulfil the inventor’s modest request, thinking all would be well. Little did he know that the entire granary of the kingdom would be emptied before each square received its quota of rice grains. Can you explain why?
Grains of rice on a chess board
Let us number the squares on the chess board from
The total number of grains of rice will be given by the formula:
Recognizing this as the sum of a geometric series with
Assuming that 50 grains of rice have a mass of one gram, the total
mass of
The moral of this story is that exponentials are beguilingly difficult for human beings to grasp. That is why logarithms and logarithmic scales, which linearize exponentials, were invented.
Napier and logarithms
Logarithms were developed by an eccentric3 Scottish laird called John Napier around 1614. He devoted twenty years of his life to achieve this. In these days of mobile phones with calculators, and computational packages on laptops, it is difficult to imagine a time when the tedium of calculations impelled people to seek methods to ease the burden.
It has been suggested that Napier got the idea for performing
additions in place of multiplications from trigonometric identities such
as
Therefore, logarithms eventually reduced multiplications to additions and exponentiations to multiplications.4 Likewise, divisions became subtractions, and taking roots was replaced by divisions. This reduction in the hierarchy of the arithmetic operations came with a commensurate reduction in computational complexity. Logarithms were indeed a great labour saving device for arithmetic operations.
Where does
Napier coined the word logarithm which means “ratio number”.
The scheme he devised was to produce a table of numbers
The strange thing is that logarithms to the base
Let us manipulate Equation 7 step-by-step
as shown below to achieve the form
Compounding of interest
Banks charge or pay compound interest on money borrowed or invested
with them. Let us assume that a sum
In point of fact, nowadays, banks do not compute interest on an
annual basis. They do so on a daily basis. Let us assume that there are
Now, what happens when the number of compounding periods grows? What happens if banks do not compute interest daily but every hour, or every minute, or every second? Is there a possible “get rich quick scheme” that involves getting paid interest every millisecond, say, or every nanosecond?
Change in compounding period
We will write a simple program to investigate how money grows as the
frequency of compounding keeps increasing. The equation we will use is
We assign
1 |
105.000000 |
2 |
105.062500 |
4 |
105.094534 |
12 |
105.116190 |
52 |
105.124584 |
365 |
105.126750 |
8760 |
105.127095 |
105.127109 |
What do you find noteworthy about this? Regardless, of how frequently
the interest is compounded, the amount
There is one trend that is apparent from the figures in the above
table, though. The numbers after the decimal place do increase
very modestly even if they seem to bounded from above by some number.
The one way to find that number is to progress from periodic compounding
to instantaneous compounding. We derive the exact value of
With the word instantaneous, we are on thin ice. Instantaneous velocity gave us calculus, with its inbuilt inconsistencies of dividing by something that is close to but not quite zero. So, we may expect something along those lines here also. Whenever instantaneous makes its presence onstage, zero and infinity cannot be far away. 😉
The road to 𝑒
There are three variables apart from steps_to_e.py
evaluates Equation 10 at logarithmic intervals
and its results are tabulated below:
n e
-----------------------------
1 2.00000000000000000
10 2.59374246010000231
100 2.70481382942152848
1000 2.71692393223559359
10000 2.71814592682492551
100000 2.71826823719229749
1000000 2.71828046909575338
10000000 2.71828169413208176
100000000 2.71828179834735773
The values are suggestive of convergence, but it is not rapid. The
limit is the historically named number 2.71828182845904524
to seventeen decimal places.
We can also countercheck with SymPy, the Python library for symbolic mathematics, by running the script below:
from sympy import *
= symbols("n")
n = limit((1 + 1 / n) ** n, n, oo)
S print(S)
to get the result E
. The script is at limit_e.py
.
The expression
What is the amount with instantaneous interest?
Instantaneous compounding does not lead to unlimited growth. We have
guessed as much from the results of evaluating Equation 9 for different values of
Now that we have defined
Thus far, we have distinguished between
We have also glancingly looked at logarithms and contrasted linear
and logarithmic scales. Central to all this is the rather diminutive
number
Hereafter, we will continue exploring
Logarithms and the hyperbola
Limits are at the heart of both the differential and integral
calculus. You have just seen one application of limits in defining the
important number
The procedure of finding the area under a closed planar curve is called quadrature or squaring. This is because the area may be thought of as being composed of little squares, which when assembled together and summed, equal the area under the curve.
Pierre de
Fermat in France had achieved great success in computing the areas
under curves of the form
Computing the area
It was one of Fermat’s contemporaries, Grégoire
Saint-Vincent, who was known as the “circle-squarer”, who found a
way to solve this problem. He also used intervals that were in a
geometric progression, but he made an important discovery in the case of
a hyperbola like
Saint-Vincent started his integration at
The account of Saint-Vincent’s method, as described below, has been drawn from several sources [6–8]. It has been simplified to use modern methods and terminology, while remaining faithful to the original in spirit and conception.
Consider Figure 6 which is Figure 5 redrawn to show how the unknown
areas
Dashed lines like
connecting the points𝑃 𝑄 and𝑃 on the arc of the hyperbola, are drawn corresponding to𝑄 and𝑥 = 1 respectively, to get a trapezium whose area,𝑥 = 𝑟 is known exactly. The area of that trapezium is used to estimate the area𝑇 1 , as explained below.𝐴 1 The point
lies on every rectangular hyperbola. Its𝑃 ( 1 , 1 ) -coordinate represents the start of both the geometric progression and the interval of integration. In our case, the common ratio𝑥 because we do not seek convergence. The initial𝑟 > 1 -value is shown as𝑥 on Figure 6.𝑟 0 = 1 also lies on the hyperbola. The straight line𝑄 ( 𝑟 , 1 𝑟 ) is an approximation to the arc𝑃 𝑄 on the hyperbola. The trapezium with heights of𝑃 𝑄 and1 and width1 𝑟 represents a first approximation to the unknown area( 𝑟 − 1 ) shown in Figure 5. The known area of the trapezium,𝐴 1 , is𝑇 1 𝑇 1 = 1 2 [ 1 1 + 1 𝑟 ] [ 𝑟 − 1 ] = 1 2 𝑟 [ 𝑟 2 − 1 ] ≈ 𝐴 1 . Moving to the next trapezium with base between
and𝑥 = 𝑟 , we have𝑥 = 𝑟 2 𝑇 2 = 1 2 [ 1 𝑟 + 1 𝑟 2 ] [ 𝑟 2 − 𝑟 ] = 𝑟 2 𝑟 2 [ 𝑟 + 1 ] [ 𝑟 − 1 ] = 1 2 𝑟 [ 𝑟 2 − 1 ] ≈ 𝐴 2 . This pattern of all the trapezium areas being the same was the remarkable observation of Saint-Vicent.
By repeatedly subdividing the intervals it may be shown that in the limit, the values of each of the
and𝑇 𝑖 will become equal. We will henceforth use𝐴 𝑖 to denote the single value shown as𝐴 ,𝐴 1 ,𝐴 2 , etc., in Figure 5, Figure 6. Note that the lower limit of area summation is𝐴 3 in all cases. We may then tabulate the respective integrals, intervals of summation, and areas so [8]:1
Integral | Upper limit | Area |
---|---|---|
0 | ||
And this is where the matter rested, until Alphonse Antonio de Sarasa—a student and later a colleague of Grégoire Saint-Vincent—took a look at the results, and realized that it was a mapping between a geometric and an arithmetic series, which meant that logarithms were involved.
A logarithm is a continuous real-valued function with the following two properties [10]:
; andl o g ( 1 ) = 0 .l o g ( 𝑎 𝑏 ) = l o g ( 𝑎 ) + l o g ( 𝑏 )
Let us see if the function
The function that equals its own derivative7
An exponential function
Let us investigate the derivative of
The next question is this: is there any value of
For finite
Solving Equation 15 for
It has been a bit of a hard slog, but we can now confidently say that
the unique function that is its own derivative and
anti-derivative is the exponential function with base
So, when we talk of the exponential function, we mean
Let us see where the foregoing leads to. Let
The natural exponential and logarithmic functions
The natural logarithm function is that logarithm function that has
We are now in a position to answer the question asked at the end of
the section Computing the area about
the base of the logarithm which gave the area under a hyperbola. The
base of the logarithm is
One may then use Equation 12 to
define the
One might wonder if there is a geometrical significance to
the number
Substituting
Logarithms and dynamic range compression
Our human senses of sight and hearing each have enormous dynamic ranges. The eye can respond to light intensities across 13 orders of magnitude.9 Likewise our ears can hear sound intensities ranging from whispers to explosions, across 12 orders of magnitude.
If you think of a weighing scale, it usually has a scale that ranges, from say 0 kg to perhaps 150 kg. Most instruments only have a limited range over which they can measure. To increase the range, you may have to switch the input to another scale before making the measurement. How then do our ears and eyes accommodate such large dynamic ranges without the need for any form of switching?
The answer lies with logarithms. Logarithms naturally compress a large linear range to a more compact one. This would be clear from the graph of the logarithm function plotted in Figure 7.
There is a “law” first propounded by the German physiologist Ernst Heinrich
Weber that the “just noticeable difference” (JND) that human beings
experienced to any physiological stimulus was related by the
differential equation
The important lesson for us is that logarithmic compression allows very large dynamic ranges to be accommodated, without input sensor switching. Logarithmic scales abound in the natural sciences and engineering: the pH scale for acidity, the Richter scale for measuring earthquake intensities, and the decibel scale for sound intensity, or for signal voltage, and power in electrical engineering, to name just a few.
Why is 𝑒 important?
We have now reached the stage where we can answer the question, “Why
is
The number
In a succeeding blog, we will see that
Acknowledgements
Thanks are due to Wolfram Alpha, Desmos, and the various AI bots, too numerous to mention, for assistance at various stages of preparation of this blog.
Feedback
Since I work independently and alone, there is every chance that unintentional mistakes have crept into this blog, due to ignorance or carelessness. Therefore, I especially appreciate your corrective and constructive feedback.
Please email me your comments and corrections.
A PDF version of this article is available for download here:
References
I later found that this link is a chapter from a draft of the book with the charmingly alliterative title Amazing and Aesthetic Aspects of Analysis [1] where it is now chapter 8.↩︎
The precursor called chaturanga was invented in India around the 600s.↩︎
This word has both a common and a mathematical meaning. Can you reconcile the two?↩︎
We touched upon this idea in the blog Varieties of Multiplication.↩︎
It is often erroneously believed that Napier used
as the base of his logarithms, but we know that his “base” was less than 1 and was indeed𝑒 .↩︎1 𝑒 These are our monomial power functions.↩︎
Beginning with the heading, this section, more than others, is heavily borrowed from Eli Maor’s excellent text e: The Story of a Number [7].↩︎
Mathematical conventions and practice might change. Programming languages might use
instead ofl o g . Beware! You have been forewarned.↩︎l n An order of magnitude conventionally means a power of ten. Two orders of magnitude thus refers to a ratio between two quantities that is either one hundred or one hundredth.↩︎