(Elements in The Philosophy of Mathematics) John Stillwell - A Concise History of Mathematics For Philosophers (2019, Cambridge University Press)
(Elements in The Philosophy of Mathematics) John Stillwell - A Concise History of Mathematics For Philosophers (2019, Cambridge University Press)
(Elements in The Philosophy of Mathematics) John Stillwell - A Concise History of Mathematics For Philosophers (2019, Cambridge University Press)
M AT H E M AT I C S
edited by
Penelope Rush
University of Tasmania
Stewart Shapiro
The Ohio State University
A CONCI SE HI STORY OF
MATHEMATI CS FOR
PHI LOSOPHERS
John Stillwell
University of San Francisco
University Printing House, Cambridge CB2 8BS, United Kingdom
314–321, 3rd Floor, Plot 3, Splendor Forum, Jasola District Centre, New Delhi – 110025,
India
www.cambridge.org
DOI: 10.1017/9781108610124
This publication is in copyright. Subject to statutory exception and to the provisions of relevant
collective licensing agreements, no reproduction of any part may take place without the written
permission of Cambridge University Press.
A catalogue record for this publication is available from the British Library.
Bibliography
A Concise History of Mathematics
for Philosophers
◈
John Stillwell
University of San Francisco
Preface
Since ancient times, there has been a struggle between mathematics and its
philosophy. As soon as there seems to be a settled view of the nature of
mathematics, some new mathematical discovery comes along to disrupt it.
Thus, the Pythagorean view that ‘all is number’ was disrupted by the
discovery of irrational lengths, and the philosophy of mathematics had to
expand to include a separate field of geometry. But this raised the question,
Can the geometric view be reconciled with the numerical view? If so, how?
And so it went, for millennia.
In many cases, advances in mathematics changed ideas about
mathematics, by forcing the acceptance of concepts previously thought
impossible or paradoxical. Thus mathematics disrupted philosophy. In the
opposite direction, philosophy kept mathematics honest by pointing out
contradictions and suggesting how concepts might be clarified in order to
resolve them. Sometimes the philosopher and the mathematician were one
and the same person – such as Descartes, Leibniz, or Bolzano – so one might
almost say that mathematics is an especially rich and stable branch of
philosophy. At any rate, if one is to understand the past and present state of
the philosophy of mathematics, one must first understand mathematics, and
its history.
The aim of the present Element is to give a brief introduction to
mathematics and its history, with particular emphasis on events that shook up
its philosophy. If you like, it is a book on ‘mathematics for philosophers’. I
try not to take a particular philosophical position, except to say that I believe
that mathematics guides philosophy, more so than the other way round. As a
corollary, I believe that mathematicians have made important contributions to
philosophy, even when it was not their intention.
Each section begins with a preview of topics to be discussed and ends
with a section highlighting the philosophical questions raised by the
mathematics. The same themes recur from section to section – intuition and
logic, meaning and existence, and the discrete and the continuous – but they
evolve under the influence of new mathematical discoveries.
Experts may be surprised that there is little or no mention of
philosophies of mathematics that were prominent in the twentieth century –
platonism, logicism, formalism, nominalism, and intuitionism, for example.
This is partly because I find none of them adequate, but mainly because I
hope to look at the philosophy of mathematics without being influenced by
labels. I want to present as much philosophically instructive mathematics as
possible and leave readers to decide how it should be sorted and labelled in
philosophical terms. My hope is that this Element will equip readers with a
‘mathematical lens’ with which to view many philosophical issues.
I thank Jeremy Avigad, Rossella Lupacchini, Wilfried Sieg, and an
anonymous referee for their helpful comments, which have resulted in many
improvements.
1 Irrational Numbers and Geometry
PREVIEW
The source of many issues in the philosophy of mathematics – the nature of
proof and truth; the meaning and existence of numbers; the role of infinity;
and the relation between geometry, algebra, and arithmetic – is Euclid’s
Elements from around 300 BCE. The Elements is best known for its
axiomatic geometry – Euclidean geometry – which includes proofs of
signature results such as the Pythagorean theorem and the existence of
exactly five regular polyhedra. However, the Elements also includes
fundamental ideas of number theory, such as the existence of infinitely many
prime numbers, the Euclidean algorithm for greatest common divisor, and (an
equivalent of) unique prime factorization.
In Euclid‘s time, as now, there was a conceptual gulf between geometry
and number theory – between measuring and counting, or between the
continuous and the discrete. The major reason for this gulf was the existence
of irrationals, discovered before Euclid’s time by the Pythagoreans and, by
the time of the Elements, the subject of a sophisticated ‘theory of
proportions’. This theory, in Book V of the Elements, made a tenuous bridge
between the continuous and the discrete. The bridge was gradually
strengthened over the centuries by the work of later mathematicians, but not
without philosophical conflicts and mathematical surprises.
These issues are the subject of this section and the next.
1.1 The Pythagorean Theorem
The Pythagorean theorem was discovered independently several times in
human history, and in several different cultures. So if any theorem typifies
mathematics – and its universality – this is it. Figure 1 illustrates the theorem:
the (grey) square on the hypotenuse of the (white) right-angled triangle is
equal to sum of the (black) squares on the other two sides.
At this point it may be noticed that and are in the same ratio
as and 1. It is not clear whether the Greeks noticed this (though they were
probably aware of something similar), but it is clear by basic algebra because
Since and are in the same ratio as and 1, applying the
Euclidean algorithm to them will produce, in two steps, yet another pair in
that ratio – and so on, forever.
Whether or not Euclid knew this particular example, he realized that the
algorithm does not terminate on a pair of lengths in irrational ratio (Book X,
Proposition 2). Thus the Euclidean algorithm elegantly separates rational
from irrational, by separating termination from non-termination; that is, finite
from infinite.
2.2 Areas and Volumes
In Book I of the Elements Euclid shows equality of various regions by adding
or subtracting equal triangles. For example, Figure 6 shows that a
parallelogram equals a rectangle of the same base and height. And Figure 7
shows that a triangle equals half a rectangle with the same base and height.
Each triangle is half the width of the triangle above it, and a calculation
shows that each group (of one, two, four, … triangles) has total area one-
fourth the area of the group above it. Thus, if the black triangle has area 1,
then the total area of the triangles is the infinite sum
If we set
then clearly
so and therefore . (This risky calculation with
infinite sums gives a quick way to guess the answer. We will see how
Archimedes did it more carefully in the next section.)
The tetrahedron is filled with prisms, as indicated in Figure 9. After two
prisms are removed (one light grey, the other darker grey), two smaller
tetrahedra remain, from which we again remove prisms, and continue.
In this way Euclid found that the volume of the tetrahedron is a similar
infinite sum,
This formula called not only for acceptance of square roots and cube roots,
but also for square roots of negative numbers – because there are equations
for which there is an obvious real solution yet .
For a long time, numbers such as were called impossible, and they
are still called ‘imaginary’. Yet they were accepted in mathematics, at least to
prove results about real numbers, because they were useful and they did not
(usually) lead to contradiction. Eventually, the system of real and imaginary
numbers came to be viewed as natural, both algebraically and geometrically.
3.1 Quadratic and Cubic Equations
As we have seen in Section 1, algebra was hamstrung in Greek mathematics
by the geometric interpretation of product. Under the geometric
interpretation, products of more than three terms have no meaning, and
products of different dimensions cannot be added. These restrictions do not
apply to the product of numbers, so in a contest between algebra for numbers
and algebra for lengths, algebra for numbers clearly wins.
This is roughly what happened in the development of algebra, which
was initially a symbolism for solving problems (especially equations) about
numbers. Until about 1600, algebra was a discipline that solved equations in
numbers but it fell back on geometry to justify its moves – because Euclid’s
Elements was still the model for mathematical proof.
The shift from geometry to an algebra of numbers began with
Diophantus, around 200 CE, in the last phase of classical Greek mathematics.
Diophantus used a symbolism that allowed products of four or more
elements. But because he was interested in finding rational solutions of
equations he used only the rational operations not the square
root operation .
The operation occurs in the general solution
(*)
(**)
and therefore
As many people have since remarked, it seems as though algebra is smarter
than we are!
At any rate, Bombelli’s example, and others like it, eventually
convinced mathematicians that it was safe to use ‘imaginary’ or ‘impossible’
numbers. Whatever means, if anything, it seems to behave like an
ordinary number and to give correct results about ordinary numbers.
3.3 The Convenience of Imaginary Numbers
In the eighteenth and nineteenth centuries mathematicians discovered many
situations in which known properties of ordinary numbers are more easily
stated, or explained, with the help of imaginary numbers. Here are some
examples.
Then the pair for the sum is the sum of the pairs for
and , and the pair for the product is likewise the product of
the pairs. It follows that any statement about sum and product of numbers of
the form is equivalent to one about real numbers, for example
Intuition and logic. The Italian algebraists at first justified the rules of
algebra (as did their Islamic predecessors) by appeal to geometric logic. But
Bombelli’s calculations with suggested that algebra had an independent
logic of its own.
Discrete and continuous. The Greek belief that only rational numbers
were really numbers gradually eroded under the influence of algebra, which
urged acceptance of square and cube roots, and of trigonometry, which urged
acceptance of the sine and cosine functions. However, there was not yet a
coherent theory of real numbers – only the belief that they could be modelled
by the points of a line.
4 Calculus and Infinitesimals
Preview
The influence of algebra grew in the seventeenth century, first in the
algebraic geometry of Fermat and Descartes, then in the infinitesimal
calculus of Newton and Leibniz. In geometry, algebra made quick work of
ancient problems about lines and conic sections and gave easy access to a
vast class of curves barely touched by the Greeks. The algebraic approach to
geometry was made possible by arithmetization of the line and plane:
identifying points of the line with real numbers and points of the plane with
pairs of real numbers. Under this identification, many curves could be
described by polynomial equations, , allowing geometric
properties to be extracted by algebraic manipulation.
Calculus extended this idea by allowing algebraic operations on
infinitesimals – quantities that behaved like non-zero numbers in calculations
but were otherwise negligible. For example, the slope of a curve
could be calculated as a quotient of infinitesimals and , where
was taken to be an infinitesimal increase in x and the corresponding
increase in .
The properties ascribed to infinitesimals were close to, if not actually,
inconsistent. Yet, like imaginary numbers, infinitesimals seemed easy and
safe to use. Mathematicians believed that, if challenged, they could reproduce
the results of infinitesimal calculus by the more rigorous method of
exhaustion. The magic of infinitesimal calculus was its ability to replace
complicated exhaustion arguments by routine calculations, so once again
convenience overcame any doubts about the existence of the mathematical
objects being used.
4.1 Infinite Series
We have now seen that Euclid and Archimedes used sums of infinite series to
find certain areas and volumes. The series in question were instances of the
infinite geometric series which has sum when
. This value can be rigorously confirmed by the method of exhaustion.
We find that the finite series
and this finite sum (for and ) is clearly less than but able to
exceed any number less than , since can be made arbitrarily small by
choosing n sufficiently large.
Therefore, the finite sums ‘exhaust’ all numbers less than and so the
infinite sum must equal .
When calculus was invented, around 1665, the geometric series was the
starting point for many other results on infinite series. However, before
calculus was invented, remarkable results about infinite series in
trigonometry were discovered in fifteenth-century India. The main
contributor to these discoveries was Madhava (c. 1340–c. 1425) and his
methods were largely algebraic. The starting point was again the geometric
series, but new series were also used ingeniously, notably the series
The latter series played a role later taken over by calculus in proving that
Madhava also discovered the series for the sine and cosine functions:
The latter series, and the related series for , were rediscovered in Europe in
the seventeenth century, and they played an important role in the
development of calculus. The independent discovery of these results in India
and Europe was perhaps the most remarkable example of the cultural
universality of mathematics since the Pythagorean theorem.
4.2 Algebraic Geometry
Infinite processes on numbers were one prerequisite for calculus. Another
was the application of algebra to geometry, or algebraic geometry for short.
The latter became possible after algebraic symbolism came to maturity in the
sixteenth century, allowing calculations with polynomials to be made just as
easily as with numbers.By the 1630s, Fermat and Descartes were able to give
an algebraic solution of a problem that is usually solved by calculus today:
finding the tangents to an algebraic curve. The setup for this problem is one
that is now familiar to high school students. Each point in the plane is
given by an ordered pair of numbers, where
so
At this stage one feels free to neglect and conclude that the slope of
for any value of x is . (In particular, when the slope is 2, so
the equation of the tangent at this point is , as found by
conventional algebra in the previous section.)
Similar calculations with and easily yield the slope of any
algebraic curve, and hence the tangent, at any point on the curve. In
particular, the slope of is . But this is just a small taste of the
magic of infinitesimals. They also allow the calculation of curved areas, such
as the area under a curve . To do this one views the area as a
function of x, taking the region between a fixed value a and a variable
value x, as in Figure 13.
since the extra strip of area has width and height that differs only
infinitesimally from . (We are again choosing a convenient moment to
neglect infinitesimals.) We conclude, dividing both sides by , that
In other words, is a function whose graph has slope is . Thus
finding areas under curves is the inverse problem to finding slopes.
If is a function we have already found as a slope , then we
can conclude that the area function is the same as , at least within a
constant. For example, if is the area under the parabola between
0 and x, then
Mathematicians will never have enough time to read all the discoveries
in Geometry (a quantity which is increasing from day to day and seems
likely in this scientific age to develop to enormous proportions) if they
continue to be presented in a rigorous form according to the manner of
the ancients.
and he and others rediscovered the sine and cosine series that had already
been discovered in India:
discovered by Euler (1748). This formula not only allows sine and cosine to
be expressed in terms of exponentials, namely
but also draws attention to their ‘hyperbolic’ analogues:
Intuition and logic. Intuition played a leading role in calculus (and still
does), where ‘infinitesimally close points on a curve’ and ‘infinitesimally thin
strips’ were used to set up equations for calculating tangents and areas. But,
once the equations were found, the force of algebraic symbolism (in the
algebra of infinitesimals) prevailed. Mathematicians believed in calculus
because of its amazing success in solving problems in geometry and
mechanics. They also thought they were on solid ground, believing that all
their results could be obtained rigorously by the method of exhaustion.
It is, I think, ‘intuitive’ that this sequence has a continuous limit curve.
But it can also be ‘seen’ that it has no tangents. It is clear from its
construction that the limit curve can be divided into four pieces
(corresponding to the four line segments in the second picture), each of which
looks exactly the same as the whole curve when magnified by 3. But if the
curve had a tangent at any point, the neighbourhood of that point would
become straighter under magnification.
Many other examples of bizarre objects that intuition can be persuaded
to grasp may be found in the book In Search of Infinity by Vilenkin (1995).
They show that our geometric intuition is capable of much more than
mathematicians thought, before the nineteenth century. Nevertheless, other
challenges to intuition arose in the nineteenth century, as we will see in the
next section.
5.5 Philosophical Issues
The proof of the intermediate value theorem is an infinite construction that
captures the point c where by repeatedly halving the interval in
which c ought to lie. It is not so different from a classical proof by exhaustion
(assuming the existence of least upper bounds), except that it depends on
being ‘given’ the function . In Bolzano’s time, it was not clear what this
meant, which created the suspicion that the intermediate value theorem is
purely an existence theorem, where the object c is claimed to exist but not
constructed.
Actually, construction of c is not a problem when is a polynomial, as
in the fundamental theorem of algebra. But more existence theorems about
continuous functions were to follow, and they became a bone of contention
with ‘constructivists’ later in the nineteenth century. In any case, the more
problematic part of Bolzano’s proof is defining the real numbers so as to
guarantee the least upper bound property and also their algebraic structure: in
short, the arithmetization of the line. This led to an eruption of unforeseen
philosophical problems, as we will see in the next two sections.
But, in the immediate wake of Bolzano and Dedekind, there were
already several issues.
Intuition and logic. Contrary to the intuition that algebra is discrete, the
fundamental theorem of algebra seems to involve continuity. And the
intuition about continuous curves (and calculus in general) demands a deeper
foundation, in an arithmetic theory of real numbers.
Meaning and existence. What does it mean to prove existence, without
giving a formula for the object claimed to exist? What precisely are the real
numbers, and what explains their completeness; that is, their closure under
various infinite operations? (Dedekind’s definition gives one answer; are
there alternatives?)
In this figure, the ‘plane’ is the interior of the disc, so its ‘points’ are
points inside the disc boundary. Its ‘lines’ are circular arcs perpendicular to
the disc boundary and ‘angles’ are the actual angles between ‘lines’. (This is
why the model is called conformal, which means that it faithfully represents
angles.) The concept of distance is via a certain formula I will not state, but
one can get a general impression of ‘distance’ as follows.
The figure shows many ‘triangles’, each of which has the same shape
because it has angles (thus, like all non-Euclidean triangles,
they have angle sum ). Since figures of the same shape have the same
size in non-Euclidean geometry, these triangles all have the same non-
Euclidean size. In particular, one sees that the non-Euclidean plane and its
‘lines’ are infinite, because each line passes through infinitely many triangles
on its way to the disc boundary. One can also guess that the ‘lines’ are curves
of shortest ‘length’, because it appears that a ‘line’ gives the shortest path
between any two points, if measured by the number of triangles it passes
through.
6.3 The Impact of Non-Euclidean Geometry
Non-Euclidean geometry finally killed any chance of deducing the parallel
axiom from Euclid’s other axioms, because the other axioms hold in
Beltrami’s model but the parallel axiom does not. One can see the failure of
the parallel axiom directly in Figure 18, where l is a ‘line’, is a point
outside l, and two ‘lines’ m and n pass through without meeting l.
Of course, there are models in which the parallel axiom does hold, so we
see that the axiom is independent of the other axioms, and it can be replaced
by the non-Euclidean parallel axiom, stating the existence of multiple
parallels, without fear of contradiction. The immediate effect of this
discovery – the first independence proof in mathematics – was increased
distrust of Euclid’s geometry as a foundation of mathematics, and increased
support for arithmetic. In fact, as we will see in the next section, an arithmetic
approach to geometry based on real vector spaces was ready to be rolled out,
though few mathematicians were ready to understand it.
Certainly, arithmetization of analysis was gaining support. Thanks to the
influence of Weierstrass, who had proposed an arithmetic definition of real
numbers independently of Dedekind in 1864, arithmetical proofs of the basic
theorems on continuous functions (and hence of the fundamental theorem of
algebra) began to circulate in the 1870s.
Another remarkable discovery, by Poincaré and Klein in the 1880s, was
that non-Euclidean geometry was already present in mathematics. They
noticed that there was non-Euclidean periodicity in the behaviour of
functions on the unit disc of complex numbers, and that the phenomenon had
been observed earlier, by Gauss, Riemann, and Schwarz, without realizing its
geometric significance. Schwarz (1872) even produced a diagram (Figure 19)
of the disc that is none other than a tessellation of the conformal disc model
by congruent non-Euclidean triangles!
It is clear from these axioms that itself is a real vector space, but a
more interesting example is , with sums and multiples of the vectors
defined, for each , by
has considerable geometric content. One can define lines, parallel lines,
and a relative concept of length along a given line. For example, one can say
that the multiples of a non-zero vector form a line through , and that
is twice as far from as . However, there is no concept of distance between
arbitrary points. To obtain the natural concept of distance, Grassmann
introduced the inner product:
This is none other than the hyperboloid, shown in Figure 20. In terms of
Minkowski distance, the geometry on this hyperboloid is none other the non-
Euclidean geometry of Beltrami! In Figure 20 (based on one due to Konrad
Polthier of the Free University of Berlin) we have indicated how triangles in
conformal disc model (Figure 17) correspond to triangles on the hyperboloid
that are equal in the sense of Minkowski distance.
Minkowski space gives substance to the wild idea of Lambert from back
in 1766 (mentioned in Section 4.5), that non-Euclidean geometry might hold
on a sphere of imaginary radius.
Euclidean and Minkowski spaces show that is a natural and
convenient foundation for geometry, both Euclidean and non-Euclidean. The
next question is: what is a foundation for ? In Section 8 we will see two
answers to this question. But first we need to take a closer look at itself, in
Section 7, to appreciate how subtle its foundation may be.
6.6 Philosophical Issues
A key step in Beltrami’s modelling of non-Euclidean geometry was his
decision to generalize the concept of length. He was emboldened to do this
by his reading of Riemann (1854), a ground-breaking work in the foundations
of geometry that was first published in 1868. In a sweeping new approach to
geometry (now known as Riemannian geometry), Riemann generalized the
arithmetized geometry of Descartes from the plane to curved spaces of any
dimension n. The points of the space, instead of being represented by ordered
pairs of numbers, were represented by ordered n-tuples . And
distance, instead of being given by the ‘Pythagorean’ formula, was obtained
by calculus from an ‘infinitesimal Pythagorean’ formula. The ‘infinitesimal
Pythagorean’ property means that the geometry of the space (as manifested
by the angle sum of a triangle, for example) approaches Euclidean geometry
in small regions, but may diverge from it in large regions. An example is the
geometry of the sphere, where small triangles have angle sum close to , but
large triangles have angle sum much greater than .
The sphere is an example of a space with constant positive curvature.
Riemann looked briefly at spaces of constant negative curvature, but it was
Beltrami who noticed that such spaces exhibit non-Euclidean geometry, and
that their ‘lines’ (curves of shortest length) can be modelled simply by
circular arcs, as in Figure 17. Thus Riemannian geometry is a thoroughgoing
arithmetization of geometry, pointing the way to an arithmetic basis for
Euclidean, non-Euclidean, and all kinds of curved geometries.
The revolution wrought by Riemann and Beltrami raised several issues
about the nature of geometry and its place in mathematics.
Intuition and logic. Models show that Euclidean and non-Euclidean
geometries are equally sound, and the real numbers provide a foundation for
both (and also for calculus and mathematical physics). It remains to find a
good foundation for the real numbers; hopefully based on the arithmetic of
natural numbers. As we know, Dedekind found one way to do this (Section
5.3).
Meaning and existence. But the real numbers involve more than
arithmetic; namely, some assumption about infinity. How much is it
necessary, and legitimate, to assume?
1 2 3 4 …
1 2 3 4 5 6 7 …
1 2 3 4 5 6 7 8 9 …
where the bottom line lists the distinct fractions in groups: first those
with , then those with , those with , and so
on.
1 2 3 4 5 6 7 …,
fails to include all the real numbers. In fact, given any list of
real numbers we can explicitly describe a real number x not on the list. The
1874 proof was not easy to follow – especially for a mathematical
community completely unprepared for it – but Cantor (1891) gave another
proof which obtains the ‘witness’ x with maximum clarity. This is the famous
(or, to some, notorious) diagonal argument.
There are many ways in which the real numbers can be
given, but to be specific we will suppose they are given as infinite decimals.
We will also ignore digits before the decimal point, so we can imagine the
numbers displayed as in Figure 21, showing just the digits after the decimal
point:
1 1 1
0 0 1
7 7 7
0 0 0
Figure 21 The diagonal construction.
Figure 21 also shows the first few digits of x. We ensure that each
by being different in the n th decimal place (and not using the digits 0 and
9 in x, because numbers with these digits can be the same even though their
digits are different – for example ). Specifically,
we define x by
Thus x is different from all the numbers , hence the given list
does not include all real numbers. Because of this, we say that the set of all
real numbers is uncountable – a countable set being one whose members can
be paired with the positive integers.
Since only countable sets can be considered ‘potentially’ infinite, the set
of real numbers is unavoidably an actual infinity.
Cantor’s argument is called ‘diagonal’ because it involves just the digits
on the diagonal of the table, shown in bold in the figure. Thus we need only
inspect a finite amount of each decimal expansion – namely, the first n digits
of – to calculate the n th digit of x. I mention this to dispel the common
misconception that the diagonal argument merely proves the existence of a
number x not on the given list . In fact it shows that x is just
as constructible as the numbers themselves.
7.3 Higher Infinities
The essence of the diagonal argument is to say: given a real number paired
with each natural number n, we can define a real number each by the
rule
since this rule makes S different from with respect to the number n.
In his 1891 paper Cantor took this train of thought to the end of the line,
showing that every set has more subsets than members. To see why,
suppose that each member x of is paired with a subset of S. But then we
can define a subset S of different from each with respect to the member
x. Namely, let
which is surely not the whole line. So the points cannot include
all real numbers. This proof can be refined to give a specific number x not
covered. We choose the first (binary) digit of x to avoid the first interval, then
the second digit to avoid the second interval, and so on – at which stage it
becomes clear that this construction is another diagonal argument.
There is in fact another route to uncountability, through Cantor’s theory
of ordinal numbers. Interesting though this is, it raises another set of issues
that we do not have space to discuss here. The existence of uncountable sets
raises enough issues on its own.
Spoilsports will no doubt object that in his lifetime Dedekind had only a
finite number of thoughts, so there must be something wrong with this
‘proof’. But who knows what the realm of Dedekind’s thoughts really is?
Dedekind’s argument had some eminent supporters, such as Russell (1903,
357).
Bolzano and Dedekind considered only countable sets, but we might
also shoot for uncountable sets. Is our apparent intuition of continuity an
intuition of uncountability and/or actual infinity?
4 added two axioms not needed for geometry but needed to derive the
properties of the real number line: the Archimedean axiom (stating that
there are no infinitesimals), and a completeness axiom (stating that there
are no gaps, in the sense of Dedekind).
• After including names 0 and S for the initial number and the successor
function , with appropriate properties, he gave inductive
definitions of sum and product:
It follows from these definitions that the functions and are defined for all
natural numbers. For example, the first equation in the definition of defines
for ; the second defines once is defined, and
hence defines for all natural numbers n, by complete induction.
Likewise, the second pair of equations defines for all n, given that is
already defined.
• The definitions of and enable all particular facts about sum and
product for the numerals to be derived, by substituting
in the defining equations. But to prove general facts, such as
, Peano provides the induction axiom for each property
: If and if for all m, then holds for each n.
(An equivalent induction axiom is that every set of natural numbers has a
least member, or that a descending sequence of natural numbers is finite. In
the latter form induction goes back to Euclid.)
Zermelo gave axioms for set theory in 1908. We omit the details, but
they are similar to the Dedekind or Peano axioms in spirit, as Zermelo
acknowledged. They assert the existence of a starting set (the empty set,
which can be viewed as 0), operations for building further sets (which,
among other things, allow successors of 0 to be built), and an axiom of
infinity stating the existence of a set including 0 and all its successors. There
is also an axiom of foundation that is similar to induction. Indeed, if the
axiom of infinity is omitted, Zermelo’s set theory has essentially the same
content as the Peano axioms. So set theory in a sense is ‘number theory
infinity’.
Set theory is an extremely powerful system, capable of covering
virtually all of mainstream mathematics. This is because it has set
construction principles – such as forming the set of all subsets of a set
– that cause explosive growth once an infinite set is present. Beginning
with Hilbert in the 1930s, there has been interest in systems with milder set
construction principles, tailored to analysis. In these systems it turns out that
we can measure the ‘strength’ of various theorems of analysis by the set
construction principles needed to prove them. (It happens surprisingly often
that we can find the ‘right set construction axioms’ to prove theorems of
analysis, rather like finding the parallel axiom to be the ‘right axiom’ to prove
many theorems of geometry. This phenomenon is studied in the new field of
reverse mathematics mentioned in Section 6.3.)
8.3 Frege’s System for Logic
So far we have been vague about how axioms may be ‘formalized’ so as to
realize Leibniz’s dream of finding truth by calculation. The missing
ingredient is formal logic. The first steps were taken by Boole (1847), in what
later became known as Boolean algebra or propositional logic. Boole noticed
that the connectives ‘or’ and ‘and’ act on propositions rather like sum and
product. They satisfy certain basic identities, such as and
from which general identities between compound propositions may
be proved by algebra.
But propositional logic is not expressive enough for mathematics, where
the internal structure of a proposition is important. Typically, a mathematical
proposition contains:
Logic symbols which include not only connectives, such as ‘and’, ‘or’,
and ‘not’, but also the quantifiers ‘for all x’ and ‘there exists an x’,
applied to any variable x.
These results, as we will show in outline below, all follow from simple
variations on Cantor’s diagonal argument.
9.1 Computability
In the seventeenth century, when Leibniz dreamed of deciding truth by
computation, the concept of computation had a rather limited meaning.
Leibniz himself had designed a computing machine capable of doing
arithmetic on numbers, and no doubt he would have accepted that algebra
was computation too. By 1850 Boole had got as far as doing propositional
logic by algebraic computation. But a general definition of computation had
to wait for the development of formal systems for mathematics, around 1900.
Only then did it become clear how broad the definition of computation
needed to be in order to make Leibniz’s dream come true.
The most influential formal system in the early twentieth century was
the Principia Mathematica of Whitehead and Russell (1910). The Principia
claimed to show how all theorems of mathematics could be generated from
particular axioms by certain rules. The rules were such that, in principle, they
were mechanical and hence could be applied without thought to strings of
symbols – eliminating all possibility of human error or bias. In the early
1920s the rules were analysed and simplified by Post, until they were reduced
to the form
We are going to show that no Turing machine S can correctly answer all
the questions . It is fair to assume that S receives question in the form
des , because can be reconstructed from des . It is also fair to
assume that S answers ‘no’ by halting on a blank square, and ‘yes’ by halting
on a non-blank square.
But then S cannot give the correct answer to question . If S, given
input , halts on a blank square, then the answer to is ‘yes’, so S should
not halt on a blank square. And if S does not halt on a blank square then the
answer to question is ‘no’, and S must halt on a blank square.
This contradiction shows that the self-examination problem cannot be
solved by Turing machine. And therefore, if the Church–Turing thesis is
correct, this problem cannot be solved by any computation whatever. As we
say, the problem is algorithmically unsolvable or, simply, unsolvable.
Since the self-examination problem is rather obviously self-defeating,
one might hope that unsolvability is an aberration, not something that
happens naturally. This is not so, because Turing machines can be
‘simulated’ by various natural systems in mathematics and logic. In fact
Church and Turing both noticed immediately that predicate logic can
‘simulate’ Turing machines, with the result the problem of deciding validity
in predicate logic – the so-called Entscheidungsproblem – is unsolvable.
9.3 Incompleteness
The notion of computability, which by some miracle seems to be complete
and absolute, stands in contrast to the notion of provability, which turns out
to be incomplete and relative. The link between the two is non-computability,
which follows from computability by the diagonal argument. The most
convenient form of the diagonal argument is the one used by Cantor (1891) to
prove that any set has more subsets than elements. Following Post, we apply
this argument to sets associated with Turing machines, called computably
enumerable sets.
A set of natural numbers is called computably enumerable if there is a
Turing machine that lists its elements. The manner of making the list is not
important, as long as any Turing machine has a computably enumerable set
associated with it (possibly the empty set), and we can observe when a given
machine lists a given number n. We will appeal to the Church–Turing
thesis to claim that any set that is intuitively listable is listable by Turing
machine. is computably enumerable, and so are many of its subsets, such
as the set of prime numbers. Moreover, we can computably enumerate all the
Turing machine descriptions, by listing them in lexicographical order. We let
be the computably enumerable set listed by the n th Turing machine.
It follows by the diagonal argument that the set
1 Suppose that is consistent and that the set of n for which proves ‘
’ is the computably enumerable set . We also assume that
is strong enough to simulate Turing machines, and hence to prove
whenever this is true. Now what can we say about the sentence
‘ ’?
2 The concept of proof in logic has been described completely (in the
case of predicate logic, the one most relevant to mathematics): there is a
computation that generates all logically valid formulas. However, there
is no algorithm that decides, given an arbitrary formula, whether that
formula is provable (hence valid) or not.
7 Despite its difficulties, the continuum has become the basis for
analysis, hence most of geometry, and most of mathematical physics.
These fields of mathematics have been arithmetized; that is, based on
axioms for the natural numbers plus certain axioms for infinite sets.
2 Perhaps it would be better to say that the only known models for
‘Cantorian’ and ‘non-Cantorian’ set theories are those constructed with the
purpose of modelling these theories. This contrasts with the situation in
geometry, where non-Euclidean geometry was found to hold in certain
structures that had been studied before there was a suitable geometric
language to describe them. We mentioned this in section 6.3.
Bibliography
Dedekind, R. (1888). Was sind und was sollen die Zahlen? Braunschweig:
Vieweg und Sohn. English translation in Essays on the Theory of Numbers,
New York: Dover, 1963.
Gödel, K. (1938). The consistency of the axiom of choice and the generalized
continuum hypothesis. Proceedings of the National Academy of Sciences 25,
220–24.
Lambert, J. H. (1766). Die Theorie der Parallellinien. Magazin für reine und
angewandte Mathematik 1786, 137–64, 325–58.
von Koch, H. (1904). Sur une courbe continue sans tangente, obtenue par une
construction géométrique élémentaire. Archiv för Matematik, Astronomi och
Fysik 1, 681–704.
Penelope Rush
University of Tasmania
From the time Penny Rush completed her thesis in the philosophy of
mathematics (2005), she has worked continuously on themes around the
realism/anti-realism divide and the nature of mathematics. Her edited
collection The Metaphysics of Logic (Cambridge University Press, 2014), and
forthcoming essay ‘Metaphysical Optimism’ (Philosophy Supplement),
highlight a particular interest in the idea of reality itself and curiosity and
respect as important philosophical methodologies.
Stewart Shapiro
The Ohio State University
Stewart Shapiro is the O’Donnell Professor of Philosophy at The Ohio State
University, a Distinguished Visiting Professor at the University of
Connecticut, and a Professorial Fellow at the University of Oslo. His major
works include Foundations without Foundationalism (1991), Philosophy of
Mathematics: Structure and Ontology (1997), Vagueness in Context (2006),
and Varieties of Logic (2014). He has taught courses in logic, philosophy of
mathematics, metaphysics, epistemology, philosophy of religion, Jewish
philosophy, social and political philosophy, and medical ethics.
About the Series