Search This Blog

Thursday, September 13, 2018

History of logic

From Wikipedia, the free encyclopedia

The history of logic deals with the study of the development of the science of valid inference (logic). Formal logics developed in ancient times in India, China, and Greece. Greek methods, particularly Aristotelian logic (or term logic) as found in the Organon, found wide application and acceptance in Western science and mathematics for millennia. The Stoics, especially Chrysippus, began the development of predicate logic.

Christian and Islamic philosophers such as Boethius (died 524) and William of Ockham (died 1347) further developed Aristotle's logic in the Middle Ages, reaching a high point in the mid-fourteenth century, with Jean Buridan. The period between the fourteenth century and the beginning of the nineteenth century saw largely decline and neglect, and at least one historian of logic regards this time as barren. Empirical methods ruled the day, as evidenced by Sir Francis Bacon's Novum Organon of 1620.

Logic revived in the mid-nineteenth century, at the beginning of a revolutionary period when the subject developed into a rigorous and formal discipline which took as its exemplar the exact method of proof used in mathematics, a hearkening back to the Greek tradition. The development of the modern "symbolic" or "mathematical" logic during this period by the likes of Boole, Frege, Russell, and Peano is the most significant in the two-thousand-year history of logic, and is arguably one of the most important and remarkable events in human intellectual history.

Progress in mathematical logic in the first few decades of the twentieth century, particularly arising from the work of Gödel and Tarski, had a significant impact on analytic philosophy and philosophical logic, particularly from the 1950s onwards, in subjects such as modal logic, temporal logic, deontic logic, and relevance logic.

Logic in the East

Logic in India

Logic began independently in ancient India and continued to develop to early modern times without any known influence from Greek logic. Medhatithi Gautama (c. 6th century BC) founded the anviksiki school of logic. The Mahabharata (12.173.45), around the 5th century BC, refers to the anviksiki and tarka schools of logic. Pāṇini (c. 5th century BC) developed a form of logic (to which Boolean logic has some similarities) for his formulation of Sanskrit grammar. Logic is described by Chanakya (c. 350-283 BC) in his Arthashastra as an independent field of inquiry.

Two of the six Indian schools of thought deal with logic: Nyaya and Vaisheshika. The Nyaya Sutras of Aksapada Gautama (c. 2nd century AD) constitute the core texts of the Nyaya school, one of the six orthodox schools of Hindu philosophy. This realist school developed a rigid five-member schema of inference involving an initial premise, a reason, an example, an application, and a conclusion. The idealist Buddhist philosophy became the chief opponent to the Naiyayikas. Nagarjuna (c. 150-250 AD), the founder of the Madhyamika ("Middle Way") developed an analysis known as the catuṣkoṭi (Sanskrit), a "four-cornered" system of argumentation that involves the systematic examination and rejection of each of the 4 possibilities of a proposition, P:
  1. P; that is, being.
  2. not P; that is, not being.
  3. P and not P; that is, being and not being.
  4. not (P or not P); that is, neither being nor not being.Under propositional logic, De Morgan's laws imply that this is equivalent to the third case (P and not P), and is therefore superfluous; there are actually only 3 cases to consider.
However, Dignaga (c 480-540 AD) is sometimes said to have developed a formal syllogism, and it was through him and his successor, Dharmakirti, that Buddhist logic reached its height; it is contested whether their analysis actually constitutes a formal syllogistic system. In particular, their analysis centered on the definition of an inference-warranting relation, "vyapti", also known as invariable concomitance or pervasion. To this end, a doctrine known as "apoha" or differentiation was developed. This involved what might be called inclusion and exclusion of defining properties.

Dignāga's famous "wheel of reason" (Hetucakra) is a method of indicating when one thing (such as smoke) can be taken as an invariable sign of another thing (like fire), but the inference is often inductive and based on past observation. Matilal remarks that Dignāga's analysis is much like John Stuart Mill's Joint Method of Agreement and Difference, which is inductive.

In addition, the traditional five-member Indian syllogism, though deductively valid, has repetitions that are unnecessary to its logical validity. As a result, some commentators see the traditional Indian syllogism as a rhetorical form that is entirely natural in many cultures of the world, and yet not as a logical form—not in the sense that all logically unnecessary elements have been omitted for the sake of analysis.

Logic in China

In China, a contemporary of Confucius, Mozi, "Master Mo", is credited with founding the Mohist school, whose canons dealt with issues relating to valid inference and the conditions of correct conclusions. In particular, one of the schools that grew out of Mohism, the Logicians, are credited by some scholars for their early investigation of formal logic. Due to the harsh rule of Legalism in the subsequent Qin Dynasty, this line of investigation disappeared in China until the introduction of Indian philosophy by Buddhists.

Logic in the West

Prehistory of logic

Valid reasoning has been employed in all periods of human history. However, logic studies the principles of valid reasoning, inference and demonstration. It is probable that the idea of demonstrating a conclusion first arose in connection with geometry, which originally meant the same as "land measurement". The ancient Egyptians discovered geometry, including the formula for the volume of a truncated pyramid. Ancient Babylon was also skilled in mathematics. Esagil-kin-apli's medical Diagnostic Handbook in the 11th century BC was based on a logical set of axioms and assumptions, while Babylonian astronomers in the 8th and 7th centuries BC employed an internal logic within their predictive planetary systems, an important contribution to the philosophy of science.

Ancient Greece before Aristotle

While the ancient Egyptians empirically discovered some truths of geometry, the great achievement of the ancient Greeks was to replace empirical methods by demonstrative proof. Both Thales and Pythagoras of the Pre-Socratic philosophers seem aware of geometry's methods.

Fragments of early proofs are preserved in the works of Plato and Aristotle, and the idea of a deductive system was probably known in the Pythagorean school and the Platonic Academy. The proofs of Euclid of Alexandria are a paradigm of Greek geometry. The three basic principles of geometry are as follows:
  • Certain propositions must be accepted as true without demonstration; such a proposition is known as an axiom of geometry.
  • Every proposition that is not an axiom of geometry must be demonstrated as following from the axioms of geometry; such a demonstration is known as a proof or a "derivation" of the proposition.
  • The proof must be formal; that is, the derivation of the proposition must be independent of the particular subject matter in question.
Further evidence that early Greek thinkers were concerned with the principles of reasoning is found in the fragment called dissoi logoi, probably written at the beginning of the fourth century BC. This is part of a protracted debate about truth and falsity. In the case of the classical Greek city-states, interest in argumentation was also stimulated by the activities of the Rhetoricians or Orators and the Sophists, who used arguments to defend or attack a thesis, both in legal and political contexts.

Thales Theorem

Thales

It is said Thales, most widely regarded as the first philosopher in the Greek tradition, measured the height of the pyramids by their shadows at the moment when his own shadow was equal to his height. Thales was said to have had a sacrifice in celebration of discovering Thales' theorem just as Pythagoras had the Pythagorean theorem.

Thales is the first known individual to use deductive reasoning applied to geometry, by deriving four corollaries to his theorem, and the first known individual to whom a mathematical discovery has been attributed. Indian and Babylonian mathematicians knew his theorem for special cases before he proved it. It is believed that Thales learned that an angle inscribed in a semicircle is a right angle during his travels to Babylon.

Pythagoras

Proof of the Pythagorean Theorem in Euclid's Elements

Before 520 BC, on one of his visits to Egypt or Greece, Pythagoras might have met the c. 54 years older Thales. The systematic study of proof seems to have begun with the school of Pythagoras (i. e. the Pythagoreans) in the late sixth century BC. Indeed, the Pythagoreans, believing all was number, are the first philosophers to emphasize form rather than matter.

Heraclitus and Parmenides

The writing of Heraclitus (c. 535 – c. 475 BC) was the first place where the word logos was given special attention in ancient Greek philosophy, Heraclitus held that everything changes and all was fire and conflicting opposites, seemingly unified only by this Logos. He is known for his obscure sayings.
This logos holds always but humans always prove unable to understand it, both before hearing it and when they have first heard it. For though all things come to be in accordance with this logos, humans are like the inexperienced when they experience such words and deeds as I set out, distinguishing each in accordance with its nature and saying how it is. But other people fail to notice what they do when awake, just as they forget what they do while asleep.
— Diels-Kranz, 22B1
Parmenides has been called the discoverer of logic.

In contrast to Heraclitus, Parmenides held that all is one and nothing changes. He may have been a dissident Pythagorean, disagreeing that One (a number) produced the many. "X is not" must always be false or meaningless. What exists can in no way not exist. Our sense perceptions with its noticing of generation and destruction are in grievous error. Instead of sense perception, Parmenides advocated logos as the means to Truth. He has been called the discoverer of logic,
For this view, that That Which Is Not exists, can never predominate. You must debar your thought from this way of search, nor let ordinary experience in its variety force you along this way, (namely, that of allowing) the eye, sightless as it is, and the ear, full of sound, and the tongue, to rule; but (you must) judge by means of the Reason (Logos) the much-contested proof which is expounded by me. (B 7.1–8.2)
Zeno of Elea, a pupil of Parmenides, had the idea of a standard argument pattern found in the method of proof known as reductio ad absurdum. This is the technique of drawing an obviously false (that is, "absurd") conclusion from an assumption, thus demonstrating that the assumption is false. Therefore, Zeno and his teacher are seen as the first to apply the art of logic. Plato's dialogue Parmenides portrays Zeno as claiming to have written a book defending the monism of Parmenides by demonstrating the absurd consequence of assuming that there is plurality. Zeno famously used this method to develop his paradoxes in his arguments against motion. Such dialectic reasoning later became popular. The members of this school were called "dialecticians" (from a Greek word meaning "to discuss").

Plato

Let no one ignorant of geometry enter here.
— Inscribed over the entrance to Plato's Academy.
Mosaic: seven men standing under a tree

None of the surviving works of the great fourth-century philosopher Plato (428–347 BC) include any formal logic, but they include important contributions to the field of philosophical logic. Plato raises three questions:
  • What is it that can properly be called true or false?
  • What is the nature of the connection between the assumptions of a valid argument and its conclusion?
  • What is the nature of definition?
The first question arises in the dialogue Theaetetus, where Plato identifies thought or opinion with talk or discourse (logos). The second question is a result of Plato's theory of Forms. Forms are not things in the ordinary sense, nor strictly ideas in the mind, but they correspond to what philosophers later called universals, namely an abstract entity common to each set of things that have the same name. In both the Republic and the Sophist, Plato suggests that the necessary connection between the assumptions of a valid argument and its conclusion corresponds to a necessary connection between "forms". The third question is about definition. Many of Plato's dialogues concern the search for a definition of some important concept (justice, truth, the Good), and it is likely that Plato was impressed by the importance of definition in mathematics. What underlies every definition is a Platonic Form, the common nature present in different particular things. Thus, a definition reflects the ultimate object of understanding, and is the foundation of all valid inference. This had a great influence on Plato's student Aristotle, in particular Aristotle's notion of the essence of a thing.

Aristotle

Aristotle

The logic of Aristotle, and particularly his theory of the syllogism, has had an enormous influence in Western thought. Aristotle was the first logician to attempt a systematic analysis of logical syntax, of noun (or term), and of verb. He was the first formal logician, in that he demonstrated the principles of reasoning by employing variables to show the underlying logical form of an argument. He sought relations of dependence which characterize necessary inference, and distinguished the validity of these relations, from the truth of the premises (the soundness of the argument). He was the first to deal with the principles of contradiction and excluded middle in a systematic way.

Front cover of book, titled "Aristotelis Logica", with an illustration of eagle on a snake
Aristotle's logic was still influential in the Renaissance

The Organon

His logical works, called the Organon, are the earliest formal study of logic that have come down to modern times. Though it is difficult to determine the dates, the probable order of writing of Aristotle's logical works is:
This diagram shows the contradictory relationships between categorical propositions in the square of opposition of Aristotelian logic.

These works are of outstanding importance in the history of logic. In the Categories, he attempts to discern all the possible things to which a term can refer; this idea underpins his philosophical work Metaphysics, which itself had a profound influence on Western thought.

He also developed a theory of non-formal logic (i.e., the theory of fallacies), which is presented in Topics and Sophistical Refutations.

On Interpretation contains a comprehensive treatment of the notions of opposition and conversion; chapter 7 is at the origin of the square of opposition (or logical square); chapter 9 contains the beginning of modal logic.

The Prior Analytics contains his exposition of the "syllogism", where three important principles are applied for the first time in history: the use of variables, a purely formal treatment, and the use of an axiomatic system.

Stoics

The other great school of Greek logic is that of the Stoics. Stoic logic traces its roots back to the late 5th century BC philosopher Euclid of Megara, a pupil of Socrates and slightly older contemporary of Plato, probably following in the tradition of Parmenides and Zeno. His pupils and successors were called "Megarians", or "Eristics", and later the "Dialecticians". The two most important dialecticians of the Megarian school were Diodorus Cronus and Philo, who were active in the late 4th century BC.

Stone bust of a bearded, grave-looking man
Chrysippus of Soli

The Stoics adopted the Megarian logic and systemized it. The most important member of the school was Chrysippus (c. 278–c. 206 BC), who was its third head, and who formalized much of Stoic doctrine. He is supposed to have written over 700 works, including at least 300 on logic, almost none of which survive. Unlike with Aristotle, we have no complete works by the Megarians or the early Stoics, and have to rely mostly on accounts (sometimes hostile) by later sources, including prominently Diogenes Laertius, Sextus Empiricus, Galen, Aulus Gellius, Alexander of Aphrodisias, and Cicero.

Three significant contributions of the Stoic school were (i) their account of modality, (ii) their theory of the Material conditional, and (iii) their account of meaning and truth.
  • Modality. According to Aristotle, the Megarians of his day claimed there was no distinction between potentiality and actuality. Diodorus Cronus defined the possible as that which either is or will be, the impossible as what will not be true, and the contingent as that which either is already, or will be false. Diodorus is also famous for what is known as his Master argument, which states that each pair of the following 3 propositions contradicts the third proposition:
  • Everything that is past is true and necessary.
  • The impossible does not follow from the possible.
  • What neither is nor will be is possible.
Diodorus used the plausibility of the first two to prove that nothing is possible if it neither is nor will be true. Chrysippus, by contrast, denied the second premise and said that the impossible could follow from the possible.
  • Conditional statements. The first logicians to debate conditional statements were Diodorus and his pupil Philo of Megara. Sextus Empiricus refers three times to a debate between Diodorus and Philo. Philo regarded a conditional as true unless it has both a true antecedent and a false consequent. Precisely, let T0 and T1 be true statements, and let F0 and F1 be false statements; then, according to Philo, each of the following conditionals is a true statement, because it is not the case that the consequent is false while the antecedent is true (it is not the case that a false statement is asserted to follow from a true statement):
  • If T0, then T1
  • If F0, then T0
  • If F0, then F1
The following conditional does not meet this requirement, and is therefore a false statement according to Philo:
  • If T0, then F0
Indeed, Sextus says "According to [Philo], there are three ways in which a conditional may be true, and one in which it may be false." Philo's criterion of truth is what would now be called a truth-functional definition of "if ... then"; it is the definition used in modern logic.
In contrast, Diodorus allowed the validity of conditionals only when the antecedent clause could never lead to an untrue conclusion. A century later, the Stoic philosopher Chrysippus attacked the assumptions of both Philo and Diodorus.
  • Meaning and truth. The most important and striking difference between Megarian-Stoic logic and Aristotelian logic is that Megarian-Stoic logic concerns propositions, not terms, and is thus closer to modern propositional logic. The Stoics distinguished between utterance (phone), which may be noise, speech (lexis), which is articulate but which may be meaningless, and discourse (logos), which is meaningful utterance. The most original part of their theory is the idea that what is expressed by a sentence, called a lekton, is something real; this corresponds to what is now called a proposition. Sextus says that according to the Stoics, three things are linked together: that which signifies, that which is signified, and the object; for example, that which signifies is the word Dion, and that which is signified is what Greeks understand but barbarians do not, and the object is Dion himself.

Medieval logic

Logic in the Middle East

Arabic text in pink and blue
A text by Avicenna, founder of Avicennian logic

The works of Al-Kindi, Al-Farabi, Avicenna, Al-Ghazali, Averroes and other Muslim logicians were based on Aristotelian logic and were important in communicating the ideas of the ancient world to the medieval West. Al-Farabi (Alfarabi) (873–950) was an Aristotelian logician who discussed the topics of future contingents, the number and relation of the categories, the relation between logic and grammar, and non-Aristotelian forms of inference. Al-Farabi also considered the theories of conditional syllogisms and analogical inference, which were part of the Stoic tradition of logic rather than the Aristotelian.

Ibn Sina (Avicenna) (980–1037) was the founder of Avicennian logic, which replaced Aristotelian logic as the dominant system of logic in the Islamic world, and also had an important influence on Western medieval writers such as Albertus Magnus. Avicenna wrote on the hypothetical syllogism and on the propositional calculus, which were both part of the Stoic logical tradition. He developed an original "temporally modalized" syllogistic theory, involving temporal logic and modal logic. He also made use of inductive logic, such as the methods of agreement, difference, and concomitant variation which are critical to the scientific method. One of Avicenna's ideas had a particularly important influence on Western logicians such as William of Ockham: Avicenna's word for a meaning or notion (ma'na), was translated by the scholastic logicians as the Latin intentio; in medieval logic and epistemology, this is a sign in the mind that naturally represents a thing. This was crucial to the development of Ockham's conceptualism: A universal term (e.g., "man") does not signify a thing existing in reality, but rather a sign in the mind (intentio in intellectu) which represents many things in reality; Ockham cites Avicenna's commentary on Metaphysics V in support of this view.

Fakhr al-Din al-Razi (b. 1149) criticised Aristotle's "first figure" and formulated an early system of inductive logic, foreshadowing the system of inductive logic developed by John Stuart Mill (1806–1873). Al-Razi's work was seen by later Islamic scholars as marking a new direction for Islamic logic, towards a Post-Avicennian logic. This was further elaborated by his student Afdaladdîn al-Khûnajî (d. 1249), who developed a form of logic revolving around the subject matter of conceptions and assents. In response to this tradition, Nasir al-Din al-Tusi (1201–1274) began a tradition of Neo-Avicennian logic which remained faithful to Avicenna's work and existed as an alternative to the more dominant Post-Avicennian school over the following centuries.

The Illuminationist school was founded by Shahab al-Din Suhrawardi (1155–1191), who developed the idea of "decisive necessity", which refers to the reduction of all modalities (necessity, possibility, contingency and impossibility) to the single mode of necessity. Ibn al-Nafis (1213–1288) wrote a book on Avicennian logic, which was a commentary of Avicenna's Al-Isharat (The Signs) and Al-Hidayah (The Guidance). Ibn Taymiyyah (1263–1328), wrote the Ar-Radd 'ala al-Mantiqiyyin, where he argued against the usefulness, though not the validity, of the syllogism and in favour of inductive reasoning. Ibn Taymiyyah also argued against the certainty of syllogistic arguments and in favour of analogy; his argument is that concepts founded on induction are themselves not certain but only probable, and thus a syllogism based on such concepts is no more certain than an argument based on analogy. He further claimed that induction itself is founded on a process of analogy. His model of analogical reasoning was based on that of juridical arguments. This model of analogy has been used in the recent work of John F. Sowa.

The Sharh al-takmil fi'l-mantiq written by Muhammad ibn Fayd Allah ibn Muhammad Amin al-Sharwani in the 15th century is the last major Arabic work on logic that has been studied. However, "thousands upon thousands of pages" on logic were written between the 14th and 19th centuries, though only a fraction of the texts written during this period have been studied by historians, hence little is known about the original work on Islamic logic produced during this later period.

Logic in medieval Europe

Top left corner of early printed text, with an illuminated S, beginning "Sicut dicit philosophus"
Brito's questions on the Old Logic

"Medieval logic" (also known as "Scholastic logic") generally means the form of Aristotelian logic developed in medieval Europe throughout roughly the period 1200–1600. For centuries after Stoic logic had been formulated, it was the dominant system of logic in the classical world. When the study of logic resumed after the Dark Ages, the main source was the work of the Christian philosopher Boethius, who was familiar with some of Aristotle's logic, but almost none of the work of the Stoics. Until the twelfth century, the only works of Aristotle available in the West were the Categories, On Interpretation, and Boethius's translation of the Isagoge of Porphyry (a commentary on the Categories). These works were known as the "Old Logic" (Logica Vetus or Ars Vetus). An important work in this tradition was the Logica Ingredientibus of Peter Abelard (1079–1142). His direct influence was small, but his influence through pupils such as John of Salisbury was great, and his method of applying rigorous logical analysis to theology shaped the way that theological criticism developed in the period that followed.

By the early thirteenth century, the remaining works of Aristotle's Organon (including the Prior Analytics, Posterior Analytics, and the Sophistical Refutations) had been recovered in the West. Logical work until then was mostly paraphrasis or commentary on the work of Aristotle. The period from the middle of the thirteenth to the middle of the fourteenth century was one of significant developments in logic, particularly in three areas which were original, with little foundation in the Aristotelian tradition that came before. These were:
  • The theory of supposition. Supposition theory deals with the way that predicates (e.g., 'man') range over a domain of individuals (e.g., all men). In the proposition 'every man is an animal', does the term 'man' range over or 'supposit for' men existing just in the present, or does the range include past and future men? Can a term supposit for a non-existing individual? Some medievalists have argued that this idea is a precursor of modern first-order logic. "The theory of supposition with the associated theories of copulatio (sign-capacity of adjectival terms), ampliatio (widening of referential domain), and distributio constitute one of the most original achievements of Western medieval logic".
  • The theory of syncategoremata. Syncategoremata are terms which are necessary for logic, but which, unlike categorematic terms, do not signify on their own behalf, but 'co-signify' with other words. Examples of syncategoremata are 'and', 'not', 'every', 'if', and so on.
  • The theory of consequences. A consequence is a hypothetical, conditional proposition: two propositions joined by the terms 'if ... then'. For example, 'if a man runs, then God exists' (Si homo currit, Deus est). A fully developed theory of consequences is given in Book III of William of Ockham's work Summa Logicae. There, Ockham distinguishes between 'material' and 'formal' consequences, which are roughly equivalent to the modern material implication and logical implication respectively. Similar accounts are given by Jean Buridan and Albert of Saxony.
The last great works in this tradition are the Logic of John Poinsot (1589–1644, known as John of St Thomas), the Metaphysical Disputations of Francisco Suarez (1548–1617), and the Logica Demonstrativa of Giovanni Girolamo Saccheri (1667–1733).

Traditional logic

The textbook tradition

Frontispiece, with title beginning "The Artes of Logike and Rethorike, plainlie set foorth in the English tounge, easie to be learned and practised".
Dudley Fenner's Art of Logic (1584)

Traditional logic generally means the textbook tradition that begins with Antoine Arnauld's and Pierre Nicole's Logic, or the Art of Thinking, better known as the Port-Royal Logic. Published in 1662, it was the most influential work on logic after Aristotle until the nineteenth century. The book presents a loosely Cartesian doctrine (that the proposition is a combining of ideas rather than terms, for example) within a framework that is broadly derived from Aristotelian and medieval term logic. Between 1664 and 1700, there were eight editions, and the book had considerable influence after that. The Port-Royal introduces the concepts of extension and intension. The account of propositions that Locke gives in the Essay is essentially that of the Port-Royal: "Verbal propositions, which are words, [are] the signs of our ideas, put together or separated in affirmative or negative sentences. So that proposition consists in the putting together or separating these signs, according as the things which they stand for agree or disagree."

Dudley Fenner helped popularize Ramist logic, a reaction against Aristotle. Another influential work was the Novum Organum by Francis Bacon, published in 1620. The title translates as "new instrument". This is a reference to Aristotle's work known as the Organon. In this work, Bacon rejects the syllogistic method of Aristotle in favor of an alternative procedure "which by slow and faithful toil gathers information from things and brings it into understanding". This method is known as inductive reasoning, a method which starts from empirical observation and proceeds to lower axioms or propositions; from these lower axioms, more general ones can be induced. For example, in finding the cause of a phenomenal nature such as heat, 3 lists should be constructed:
  • The presence list: a list of every situation where heat is found.
  • The absence list: a list of every situation that is similar to at least one of those of the presence list, except for the lack of heat.
  • The variability list: a list of every situation where heat can vary.
Then, the form nature (or cause) of heat may be defined as that which is common to every situation of the presence list, and which is lacking from every situation of the absence list, and which varies by degree in every situation of the variability list.

Other works in the textbook tradition include Isaac Watts's Logick: Or, the Right Use of Reason (1725), Richard Whately's Logic (1826), and John Stuart Mill's A System of Logic (1843). Although the latter was one of the last great works in the tradition, Mill's view that the foundations of logic lie in introspection influenced the view that logic is best understood as a branch of psychology, a view which dominated the next fifty years of its development, especially in Germany.

Logic in Hegel's philosophy

Georg Wilhelm Friedrich Hegel

G.W.F. Hegel indicated the importance of logic to his philosophical system when he condensed his extensive Science of Logic into a shorter work published in 1817 as the first volume of his Encyclopaedia of the Philosophical Sciences. The "Shorter" or "Encyclopaedia" Logic, as it is often known, lays out a series of transitions which leads from the most empty and abstract of categories—Hegel begins with "Pure Being" and "Pure Nothing"—to the "Absolute, the category which contains and resolves all the categories which preceded it. Despite the title, Hegel's Logic is not really a contribution to the science of valid inference. Rather than deriving conclusions about concepts through valid inference from premises, Hegel seeks to show that thinking about one concept compels thinking about another concept (one cannot, he argues, possess the concept of "Quality" without the concept of "Quantity"); this compulsion is, supposedly, not a matter of individual psychology, because it arises almost organically from the content of the concepts themselves. His purpose is to show the rational structure of the "Absolute"—indeed of rationality itself. The method by which thought is driven from one concept to its contrary, and then to further concepts, is known as the Hegelian dialectic.

Although Hegel's Logic has had little impact on mainstream logical studies, its influence can be seen elsewhere:
  • Carl von Prantl's Geschichte der Logik in Abendland (1855–1867).
  • The work of the British Idealists, such as F.H. Bradley's Principles of Logic (1883).
  • The economic, political, and philosophical studies of Karl Marx, and in the various schools of Marxism.

Logic and psychology

Between the work of Mill and Frege stretched half a century during which logic was widely treated as a descriptive science, an empirical study of the structure of reasoning, and thus essentially as a branch of psychology. The German psychologist Wilhelm Wundt, for example, discussed deriving "the logical from the psychological laws of thought", emphasizing that "psychological thinking is always the more comprehensive form of thinking." This view was widespread among German philosophers of the period:
  • Theodor Lipps described logic as "a specific discipline of psychology".
  • Christoph von Sigwart understood logical necessity as grounded in the individual's compulsion to think in a certain way.
  • Benno Erdmann argued that "logical laws only hold within the limits of our thinking".
Such was the dominant view of logic in the years following Mill's work. This psychological approach to logic was rejected by Gottlob Frege. It was also subjected to an extended and destructive critique by Edmund Husserl in the first volume of his Logical Investigations (1900), an assault which has been described as "overwhelming". Husserl argued forcefully that grounding logic in psychological observations implied that all logical truths remained unproven, and that skepticism and relativism were unavoidable consequences.

Such criticisms did not immediately extirpate what is called "psychologism". For example, the American philosopher Josiah Royce, while acknowledging the force of Husserl's critique, remained "unable to doubt" that progress in psychology would be accompanied by progress in logic, and vice versa.

Rise of modern logic

The period between the fourteenth century and the beginning of the nineteenth century had been largely one of decline and neglect, and is generally regarded as barren by historians of logic. The revival of logic occurred in the mid-nineteenth century, at the beginning of a revolutionary period where the subject developed into a rigorous and formalistic discipline whose exemplar was the exact method of proof used in mathematics. The development of the modern "symbolic" or "mathematical" logic during this period is the most significant in the 2000-year history of logic, and is arguably one of the most important and remarkable events in human intellectual history.

A number of features distinguish modern logic from the old Aristotelian or traditional logic, the most important of which are as follows: Modern logic is fundamentally a calculus whose rules of operation are determined only by the shape and not by the meaning of the symbols it employs, as in mathematics. Many logicians were impressed by the "success" of mathematics, in that there had been no prolonged dispute about any truly mathematical result. C.S. Peirce noted that even though a mistake in the evaluation of a definite integral by Laplace led to an error concerning the moon's orbit that persisted for nearly 50 years, the mistake, once spotted, was corrected without any serious dispute. Peirce contrasted this with the disputation and uncertainty surrounding traditional logic, and especially reasoning in metaphysics. He argued that a truly "exact" logic would depend upon mathematical, i.e., "diagrammatic" or "iconic" thought. "Those who follow such methods will ... escape all error except such as will be speedily corrected after it is once suspected". Modern logic is also "constructive" rather than "abstractive"; i.e., rather than abstracting and formalising theorems derived from ordinary language (or from psychological intuitions about validity), it constructs theorems by formal methods, then looks for an interpretation in ordinary language. It is entirely symbolic, meaning that even the logical constants (which the medieval logicians called "syncategoremata") and the categoric terms are expressed in symbols.

Modern logic

The development of modern logic falls into roughly five periods:
  • The embryonic period from Leibniz to 1847, when the notion of a logical calculus was discussed and developed, particularly by Leibniz, but no schools were formed, and isolated periodic attempts were abandoned or went unnoticed.
  • The algebraic period from Boole's Analysis to Schröder's Vorlesungen. In this period, there were more practitioners, and a greater continuity of development.
  • The logicist period from the Begriffsschrift of Frege to the Principia Mathematica of Russell and Whitehead. The aim of the "logicist school" was to incorporate the logic of all mathematical and scientific discourse in a single unified system which, taking as a fundamental principle that all mathematical truths are logical, did not accept any non-logical terminology. The major logicists were Frege, Russell, and the early Wittgenstein. It culminates with the Principia, an important work which includes a thorough examination and attempted solution of the antinomies which had been an obstacle to earlier progress.
  • The metamathematical period from 1910 to the 1930s, which saw the development of metalogic, in the finitist system of Hilbert, and the non-finitist system of Löwenheim and Skolem, the combination of logic and metalogic in the work of Gödel and Tarski. Gödel's incompleteness theorem of 1931 was one of the greatest achievements in the history of logic. Later in the 1930s, Gödel developed the notion of set-theoretic constructibility.
  • The period after World War II, when mathematical logic branched into four inter-related but separate areas of research: model theory, proof theory, computability theory, and set theory, and its ideas and methods began to influence philosophy.

Embryonic period

Leibniz

The idea that inference could be represented by a purely mechanical process is found as early as Raymond Llull, who proposed a (somewhat eccentric) method of drawing conclusions by a system of concentric rings. The work of logicians such as the Oxford Calculators led to a method of using letters instead of writing out logical calculations (calculationes) in words, a method used, for instance, in the Logica magna by Paul of Venice. Three hundred years after Llull, the English philosopher and logician Thomas Hobbes suggested that all logic and reasoning could be reduced to the mathematical operations of addition and subtraction. The same idea is found in the work of Leibniz, who had read both Llull and Hobbes, and who argued that logic can be represented through a combinatorial process or calculus. But, like Llull and Hobbes, he failed to develop a detailed or comprehensive system, and his work on this topic was not published until long after his death. Leibniz says that ordinary languages are subject to "countless ambiguities" and are unsuited for a calculus, whose task is to expose mistakes in inference arising from the forms and structures of words;[103] hence, he proposed to identify an alphabet of human thought comprising fundamental concepts which could be composed to express complex ideas, and create a calculus ratiocinator that would make all arguments "as tangible as those of the Mathematicians, so that we can find our error at a glance, and when there are disputes among persons, we can simply say: Let us calculate."

Gergonne (1816) said that reasoning does not have to be about objects about which one has perfectly clear ideas, because algebraic operations can be carried out without having any idea of the meaning of the symbols involved. Bolzano anticipated a fundamental idea of modern proof theory when he defined logical consequence or "deducibility" in terms of variables:
Hence I say that propositions M, N, O,… are deducible from propositions A, B, C, D,… with respect to variable parts i, j,…, if every class of ideas whose substitution for i, j,… makes all of A, B, C, D,… true, also makes all of M, N, O,… true. Occasionally, since it is customary, I shall say that propositions M, N, O,… follow, or can be inferred or derived, from A, B, C, D,…. Propositions A, B, C, D,… I shall call the premises, M, N, O,… the conclusions.
This is now known as semantic validity.

Algebraic period

George Boole

Modern logic begins with what is known as the "algebraic school", originating with Boole and including Peirce, Jevons, Schröder, and Venn. Their objective was to develop a calculus to formalise reasoning in the area of classes, propositions, and probabilities. The school begins with Boole's seminal work Mathematical Analysis of Logic which appeared in 1847, although De Morgan (1847) is its immediate precursor. The fundamental idea of Boole's system is that algebraic formulae can be used to express logical relations. This idea occurred to Boole in his teenage years, working as an usher in a private school in Lincoln, Lincolnshire. For example, let x and y stand for classes let the symbol = signify that the classes have the same members, xy stand for the class containing all and only the members of x and y and so on. Boole calls these elective symbols, i.e. symbols which select certain objects for consideration. An expression in which elective symbols are used is called an elective function, and an equation of which the members are elective functions, is an elective equation. The theory of elective functions and their "development" is essentially the modern idea of truth-functions and their expression in disjunctive normal form.

Boole's system admits of two interpretations, in class logic, and propositional logic. Boole distinguished between "primary propositions" which are the subject of syllogistic theory, and "secondary propositions", which are the subject of propositional logic, and showed how under different "interpretations" the same algebraic system could represent both. An example of a primary proposition is "All inhabitants are either Europeans or Asiatics." An example of a secondary proposition is "Either all inhabitants are Europeans or they are all Asiatics." These are easily distinguished in modern propositional calculus, where it is also possible to show that the first follows from the second, but it is a significant disadvantage that there is no way of representing this in the Boolean system.

In his Symbolic Logic (1881), John Venn used diagrams of overlapping areas to express Boolean relations between classes or truth-conditions of propositions. In 1869 Jevons realised that Boole's methods could be mechanised, and constructed a "logical machine" which he showed to the Royal Society the following year. In 1885 Allan Marquand proposed an electrical version of the machine that is still extant (picture at the Firestone Library).

Charles Sanders Peirce

The defects in Boole's system (such as the use of the letter v for existential propositions) were all remedied by his followers. Jevons published Pure Logic, or the Logic of Quality apart from Quantity in 1864, where he suggested a symbol to signify exclusive or, which allowed Boole's system to be greatly simplified. This was usefully exploited by Schröder when he set out theorems in parallel columns in his Vorlesungen (1890–1905). Peirce (1880) showed how all the Boolean elective functions could be expressed by the use of a single primitive binary operation, "neither ... nor ..." and equally well "not both ... and ...", however, like many of Peirce's innovations, this remained unknown or unnoticed until Sheffer rediscovered it in 1913. Boole's early work also lacks the idea of the logical sum which originates in Peirce (1867), Schröder (1877) and Jevons (1890), and the concept of inclusion, first suggested by Gergonne (1816) and clearly articulated by Peirce (1870).

Coloured diagram of 4 interlocking sets
Boolean multiples

The success of Boole's algebraic system suggested that all logic must be capable of algebraic representation, and there were attempts to express a logic of relations in such form, of which the most ambitious was Schröder's monumental Vorlesungen über die Algebra der Logik ("Lectures on the Algebra of Logic", vol iii 1895), although the original idea was again anticipated by Peirce.

Boole's unwavering acceptance of Aristotle's logic is emphasized by the historian of logic John Corcoran in an accessible introduction to Laws of Thought Corcoran also wrote a point-by-point comparison of Prior Analytics and Laws of Thought. According to Corcoran, Boole fully accepted and endorsed Aristotle's logic. Boole's goals were "to go under, over, and beyond" Aristotle's logic by 1) providing it with mathematical foundations involving equations, 2) extending the class of problems it could treat — from assessing validity to solving equations — and 3) expanding the range of applications it could handle — e.g. from propositions having only two terms to those having arbitrarily many.

More specifically, Boole agreed with what Aristotle said; Boole's 'disagreements', if they might be called that, concern what Aristotle did not say. First, in the realm of foundations, Boole reduced the four propositional forms of Aristotelian logic to formulas in the form of equations — by itself a revolutionary idea. Second, in the realm of logic's problems, Boole's addition of equation solving to logic — another revolutionary idea — involved Boole's doctrine that Aristotle's rules of inference (the "perfect syllogisms") must be supplemented by rules for equation solving. Third, in the realm of applications, Boole's system could handle multi-term propositions and arguments whereas Aristotle could handle only two-termed subject-predicate propositions and arguments. For example, Aristotle's system could not deduce "No quadrangle that is a square is a rectangle that is a rhombus" from "No square that is a quadrangle is a rhombus that is a rectangle" or from "No rhombus that is a rectangle is a square that is a quadrangle".

Logicist period

Gottlob Frege.

After Boole, the next great advances were made by the German mathematician Gottlob Frege. Frege's objective was the program of Logicism, i.e. demonstrating that arithmetic is identical with logic. Frege went much further than any of his predecessors in his rigorous and formal approach to logic, and his calculus or Begriffsschrift is important. Frege also tried to show that the concept of number can be defined by purely logical means, so that (if he was right) logic includes arithmetic and all branches of mathematics that are reducible to arithmetic. He was not the first writer to suggest this. In his pioneering work Die Grundlagen der Arithmetik (The Foundations of Arithmetic), sections 15–17, he acknowledges the efforts of Leibniz, J.S. Mill as well as Jevons, citing the latter's claim that "algebra is a highly developed logic, and number but logical discrimination."

Frege's first work, the Begriffsschrift ("concept script") is a rigorously axiomatised system of propositional logic, relying on just two connectives (negational and conditional), two rules of inference (modus ponens and substitution), and six axioms. Frege referred to the "completeness" of this system, but was unable to prove this. The most significant innovation, however, was his explanation of the quantifier in terms of mathematical functions. Traditional logic regards the sentence "Caesar is a man" as of fundamentally the same form as "all men are mortal." Sentences with a proper name subject were regarded as universal in character, interpretable as "every Caesar is a man". At the outset Frege abandons the traditional "concepts subject and predicate", replacing them with argument and function respectively, which he believes "will stand the test of time. It is easy to see how regarding a content as a function of an argument leads to the formation of concepts. Furthermore, the demonstration of the connection between the meanings of the words if, and, not, or, there is, some, all, and so forth, deserves attention". Frege argued that the quantifier expression "all men" does not have the same logical or semantic form as "all men", and that the universal proposition "every A is B" is a complex proposition involving two functions, namely ' – is A' and ' – is B' such that whatever satisfies the first, also satisfies the second. In modern notation, this would be expressed as
{\displaystyle \forall \;x{\big (}A(x)\rightarrow B(x){\big )}}
In English, "for all x, if Ax then Bx". Thus only singular propositions are of subject-predicate form, and they are irreducibly singular, i.e. not reducible to a general proposition. Universal and particular propositions, by contrast, are not of simple subject-predicate form at all. If "all mammals" were the logical subject of the sentence "all mammals are land-dwellers", then to negate the whole sentence we would have to negate the predicate to give "all mammals are not land-dwellers". But this is not the case. This functional analysis of ordinary-language sentences later had a great impact on philosophy and linguistics.

This means that in Frege's calculus, Boole's "primary" propositions can be represented in a different way from "secondary" propositions. "All inhabitants are either men or women" is

Straight line with bend; text "x" over bend; text "F(x)" to the right of the line.
 
Frege's "Concept Script"
{\displaystyle \forall \;x{\Big (}I(x)\rightarrow {\big (}M(x)\lor W(x){\big )}{\Big )}}
whereas "All the inhabitants are men or all the inhabitants are women" is
{\displaystyle \forall \;x{\big (}I(x)\rightarrow M(x){\big )}\lor \forall \;x{\big (}I(x)\rightarrow W(x){\big )}}
As Frege remarked in a critique of Boole's calculus:
"The real difference is that I avoid [the Boolean] division into two parts ... and give a homogeneous presentation of the lot. In Boole the two parts run alongside one another, so that one is like the mirror image of the other, but for that very reason stands in no organic relation to it'
As well as providing a unified and comprehensive system of logic, Frege's calculus also resolved the ancient problem of multiple generality. The ambiguity of "every girl kissed a boy" is difficult to express in traditional logic, but Frege's logic resolves this through the different scope of the quantifiers. Thus
{\displaystyle \forall \;x{\Big (}G(x)\rightarrow \exists \;y{\big (}B(y)\land K(x,y){\big )}{\Big )}}
Peano

means that to every girl there corresponds some boy (any one will do) who the girl kissed. But
{\displaystyle \exists \;x{\Big (}B(x)\land \forall \;y{\big (}G(y)\rightarrow K(y,x){\big )}{\Big )}}
means that there is some particular boy whom every girl kissed. Without this device, the project of logicism would have been doubtful or impossible. Using it, Frege provided a definition of the ancestral relation, of the many-to-one relation, and of mathematical induction.

Ernst Zermelo

This period overlaps with the work of what is known as the "mathematical school", which included Dedekind, Pasch, Peano, Hilbert, Zermelo, Huntington, Veblen and Heyting. Their objective was the axiomatisation of branches of mathematics like geometry, arithmetic, analysis and set theory. Most notable was Hilbert's Program, which sought to ground all of mathematics to a finite set of axioms, proving its consistency by "finitistic" means and providing a procedure which would decide the truth or falsity of any mathematical statement. The standard axiomatization of the natural numbers is named the Peano axioms in his honor. Peano maintained a clear distinction between mathematical and logical symbols. While unaware of Frege's work, he independently recreated his logical apparatus based on the work of Boole and Schröder.

The logicist project received a near-fatal setback with the discovery of a paradox in 1901 by Bertrand Russell. This proved Frege's naive set theory led to a contradiction. Frege's theory contained the axiom that for any formal criterion, there is a set of all objects that meet the criterion. Russell showed that a set containing exactly the sets that are not members of themselves would contradict its own definition (if it is not a member of itself, it is a member of itself, and if it is a member of itself, it is not). This contradiction is now known as Russell's paradox. One important method of resolving this paradox was proposed by Ernst Zermelo. Zermelo set theory was the first axiomatic set theory. It was developed into the now-canonical Zermelo–Fraenkel set theory (ZF). Russell's paradox symbolically is as follows:
{\text{Let }}R=\{x\mid x\not \in x\}{\text{, then }}R\in R\iff R\not \in R
The monumental Principia Mathematica, a three-volume work on the foundations of mathematics, written by Russell and Alfred North Whitehead and published 1910–13 also included an attempt to resolve the paradox, by means of an elaborate system of types: a set of elements is of a different type than is each of its elements (set is not the element; one element is not the set) and one cannot speak of the "set of all sets". The Principia was an attempt to derive all mathematical truths from a well-defined set of axioms and inference rules in symbolic logic.

Metamathematical period

Kurt Gödel

The names of Gödel and Tarski dominate the 1930s, a crucial period in the development of metamathematics – the study of mathematics using mathematical methods to produce metatheories, or mathematical theories about other mathematical theories. Early investigations into metamathematics had been driven by Hilbert's program. Work on metamathematics culminated in the work of Gödel, who in 1929 showed that a given first-order sentence is deducible if and only if it is logically valid – i.e. it is true in every structure for its language. This is known as Gödel's completeness theorem. A year later, he proved two important theorems, which showed Hibert's program to be unattainable in its original form. The first is that no consistent system of axioms whose theorems can be listed by an effective procedure such as an algorithm or computer program is capable of proving all facts about the natural numbers. For any such system, there will always be statements about the natural numbers that are true, but that are unprovable within the system. The second is that if such a system is also capable of proving certain basic facts about the natural numbers, then the system cannot prove the consistency of the system itself. These two results are known as Gödel's incompleteness theorems, or simply Gödel's Theorem. Later in the decade, Gödel developed the concept of set-theoretic constructibility, as part of his proof that the axiom of choice and the continuum hypothesis are consistent with Zermelo–Fraenkel set theory. In proof theory, Gerhard Gentzen developed natural deduction and the sequent calculus. The former attempts to model logical reasoning as it 'naturally' occurs in practice and is most easily applied to intuitionistic logic, while the latter was devised to clarify the derivation of logical proofs in any formal system. Since Gentzen's work, natural deduction and sequent calculi have been widely applied in the fields of proof theory, mathematical logic and computer science. Gentzen also proved normalization and cut-elimination theorems for intuitionistic and classical logic which could be used to reduce logical proofs to a normal form.

Balding man, with bookshelf in background
Alfred Tarski

Alfred Tarski, a pupil of Łukasiewicz, is best known for his definition of truth and logical consequence, and the semantic concept of logical satisfaction. In 1933, he published (in Polish) The concept of truth in formalized languages, in which he proposed his semantic theory of truth: a sentence such as "snow is white" is true if and only if snow is white. Tarski's theory separated the metalanguage, which makes the statement about truth, from the object language, which contains the sentence whose truth is being asserted, and gave a correspondence (the T-schema) between phrases in the object language and elements of an interpretation. Tarski's approach to the difficult idea of explaining truth has been enduringly influential in logic and philosophy, especially in the development of model theory. Tarski also produced important work on the methodology of deductive systems, and on fundamental principles such as completeness, decidability, consistency and definability. According to Anita Feferman, Tarski "changed the face of logic in the twentieth century".

Alonzo Church and Alan Turing proposed formal models of computability, giving independent negative solutions to Hilbert's Entscheidungsproblem in 1936 and 1937, respectively. The Entscheidungsproblem asked for a procedure that, given any formal mathematical statement, would algorithmically determine whether the statement is true. Church and Turing proved there is no such procedure; Turing's paper introduced the halting problem as a key example of a mathematical problem without an algorithmic solution.

Church's system for computation developed into the modern λ-calculus, while the Turing machine became a standard model for a general-purpose computing device. It was soon shown that many other proposed models of computation were equivalent in power to those proposed by Church and Turing. These results led to the Church–Turing thesis that any deterministic algorithm that can be carried out by a human can be carried out by a Turing machine. Church proved additional undecidability results, showing that both Peano arithmetic and first-order logic are undecidable. Later work by Emil Post and Stephen Cole Kleene in the 1940s extended the scope of computability theory and introduced the concept of degrees of unsolvability.

The results of the first few decades of the twentieth century also had an impact upon analytic philosophy and philosophical logic, particularly from the 1950s onwards, in subjects such as modal logic, temporal logic, deontic logic, and relevance logic.

Logic after WWII

Man with a beard and straw hat on a beach

After World War II, mathematical logic branched into four inter-related but separate areas of research: model theory, proof theory, computability theory, and set theory.

In set theory, the method of forcing revolutionized the field by providing a robust method for constructing models and obtaining independence results. Paul Cohen introduced this method in 1963 to prove the independence of the continuum hypothesis and the axiom of choice from Zermelo–Fraenkel set theory. His technique, which was simplified and extended soon after its introduction, has since been applied to many other problems in all areas of mathematical logic.

Computability theory had its roots in the work of Turing, Church, Kleene, and Post in the 1930s and 40s. It developed into a study of abstract computability, which became known as recursion theory. The priority method, discovered independently by Albert Muchnik and Richard Friedberg in the 1950s, led to major advances in the understanding of the degrees of unsolvability and related structures. Research into higher-order computability theory demonstrated its connections to set theory. The fields of constructive analysis and computable analysis were developed to study the effective content of classical mathematical theorems; these in turn inspired the program of reverse mathematics. A separate branch of computability theory, computational complexity theory, was also characterized in logical terms as a result of investigations into descriptive complexity.

Model theory applies the methods of mathematical logic to study models of particular mathematical theories. Alfred Tarski published much pioneering work in the field, which is named after a series of papers he published under the title Contributions to the theory of models. In the 1960s, Abraham Robinson used model-theoretic techniques to develop calculus and analysis based on infinitesimals, a problem that first had been proposed by Leibniz.

In proof theory, the relationship between classical mathematics and intuitionistic mathematics was clarified via tools such as the realizability method invented by Georg Kreisel and Gödel's Dialectica interpretation. This work inspired the contemporary area of proof mining. The Curry-Howard correspondence emerged as a deep analogy between logic and computation, including a correspondence between systems of natural deduction and typed lambda calculi used in computer science. As a result, research into this class of formal systems began to address both logical and computational aspects; this area of research came to be known as modern type theory. Advances were also made in ordinal analysis and the study of independence results in arithmetic such as the Paris–Harrington theorem.

This was also a period, particularly in the 1950s and afterwards, when the ideas of mathematical logic begin to influence philosophical thinking. For example, tense logic is a formalised system for representing, and reasoning about, propositions qualified in terms of time. The philosopher Arthur Prior played a significant role in its development in the 1960s. Modal logics extend the scope of formal logic to include the elements of modality (for example, possibility and necessity). The ideas of Saul Kripke, particularly about possible worlds, and the formal system now called Kripke semantics have had a profound impact on analytic philosophy. His best known and most influential work is Naming and Necessity (1980). Deontic logics are closely related to modal logics: they attempt to capture the logical features of obligation, permission and related concepts. Although some basic novelties syncretizing mathematical and philosophical logic were shown by Bolzano in the early 1800s, it was Ernst Mally, a pupil of Alexius Meinong, who was to propose the first formal deontic system in his Grundgesetze des Sollens, based on the syntax of Whitehead's and Russell's propositional calculus.

Another logical system founded after World War II was fuzzy logic by Azerbaijani mathematician Lotfi Asker Zadeh in 1965.

Mathematical proof

From Wikipedia, the free encyclopedia
 
P. Oxy. 29, one of the oldest surviving fragments of Euclid's Elements, a textbook used for millennia to teach proof-writing techniques. The diagram accompanies Book II, Proposition 5.
 
In mathematics, a proof is an inferential argument for a mathematical statement. In the argument, other previously established statements, such as theorems, can be used. In principle, a proof can be traced back to self-evident or assumed statements, known as axioms, along with accepted rules of inference. Axioms may be treated as conditions that must be met before the statement applies. Proofs are examples of exhaustive deductive reasoning or inductive reasoning and are distinguished from empirical arguments or non-exhaustive inductive reasoning (or "reasonable expectation"). A proof must demonstrate that a statement is always true (occasionally by listing all possible cases and showing that it holds in each), rather than enumerate many confirmatory cases. An unproved proposition that is believed to be true is known as a conjecture.

Proofs employ logic but usually include some amount of natural language which usually admits some ambiguity. In fact, the vast majority of proofs in written mathematics can be considered as applications of rigorous informal logic. Purely formal proofs, written in symbolic language instead of natural language, are considered in proof theory. The distinction between formal and informal proofs has led to much examination of current and historical mathematical practice, quasi-empiricism in mathematics, and so-called folk mathematics (in both senses of that term). The philosophy of mathematics is concerned with the role of language and logic in proofs, and mathematics as a language.

History and etymology

The word "proof" comes from the Latin probare meaning "to test". Related modern words are the English "probe", "probation", and "probability", the Spanish probar (to smell or taste, or (lesser use) touch or test), Italian provare (to try), and the German probieren (to try). The early use of "probity" was in the presentation of legal evidence. A person of authority, such as a nobleman, was said to have probity, whereby the evidence was by his relative authority, which outweighed empirical testimony.

Plausibility arguments using heuristic devices such as pictures and analogies preceded strict mathematical proof. It is likely that the idea of demonstrating a conclusion first arose in connection with geometry, which originally meant the same as "land measurement". The development of mathematical proof is primarily the product of ancient Greek mathematics, and one of the greatest achievements thereof. Thales (624–546 BCE) and Hippocrates of Chios (c470-410 BCE) proved some theorems in geometry. Eudoxus (408–355 BCE) and Theaetetus (417–369 BCE) formulated theorems but did not prove them. Aristotle (384–322 BCE) said definitions should describe the concept being defined in terms of other concepts already known. Mathematical proofs were revolutionized by Euclid (300 BCE), who introduced the axiomatic method still in use today, starting with undefined terms and axioms (propositions regarding the undefined terms assumed to be self-evidently true from the Greek "axios" meaning "something worthy"), and used these to prove theorems using deductive logic. His book, the Elements, was read by anyone who was considered educated in the West until the middle of the 20th century. In addition to theorems of geometry, such as the Pythagorean theorem, the Elements also covers number theory, including a proof that the square root of two is irrational and that there are infinitely many prime numbers.

Further advances took place in medieval Islamic mathematics. While earlier Greek proofs were largely geometric demonstrations, the development of arithmetic and algebra by Islamic mathematicians allowed more general proofs that no longer depended on geometry. In the 10th century CE, the Iraqi mathematician Al-Hashimi provided general proofs for numbers (rather than geometric demonstrations) as he considered multiplication, division, etc. for "lines." He used this method to provide a proof of the existence of irrational numbers. An inductive proof for arithmetic sequences was introduced in the Al-Fakhri (1000) by Al-Karaji, who used it to prove the binomial theorem and properties of Pascal's triangle. Alhazen also developed the method of proof by contradiction, as the first attempt at proving the Euclidean parallel postulate.

Modern proof theory treats proofs as inductively defined data structures. There is no longer an assumption that axioms are "true" in any sense; this allows for parallel mathematical theories built on alternate sets of axioms (see Axiomatic set theory and Non-Euclidean geometry for examples).

Nature and purpose

As practiced, a proof is expressed in natural language and is a rigorous argument intended to convince the audience of the truth of a statement. The standard of rigor is not absolute and has varied throughout history. A proof can be presented differently depending on the intended audience. In order to gain acceptance, a proof has to meet communal statements of rigor; an argument considered vague or incomplete may be rejected.

The concept of a proof is formalized in the field of mathematical logic. A formal proof is written in a formal language instead of a natural language. A formal proof is defined as sequence of formulas in a formal language, in which each formula is a logical consequence of preceding formulas. Having a definition of formal proof makes the concept of proof amenable to study. Indeed, the field of proof theory studies formal proofs and their properties, for example, the property that a statement has a formal proof. An application of proof theory is to show that certain undecidable statements are not provable.

The definition of a formal proof is intended to capture the concept of proofs as written in the practice of mathematics. The soundness of this definition amounts to the belief that a published proof can, in principle, be converted into a formal proof. However, outside the field of automated proof assistants, this is rarely done in practice. A classic question in philosophy asks whether mathematical proofs are analytic or synthetic. Kant, who introduced the analytic-synthetic distinction, believed mathematical proofs are synthetic.

Proofs may be viewed as aesthetic objects, admired for their mathematical beauty. The mathematician Paul Erdős was known for describing proofs he found particularly elegant as coming from "The Book", a hypothetical tome containing the most beautiful method(s) of proving each theorem. The book Proofs from THE BOOK, published in 2003, is devoted to presenting 32 proofs its editors find particularly pleasing.

Methods

Direct proof

In direct proof, the conclusion is established by logically combining the axioms, definitions, and earlier theorems. For example, direct proof can be used to establish that the sum of two even integers is always even:
Consider two even integers x and y. Since they are even, they can be written as x = 2a and y = 2b, respectively, for integers a and b. Then the sum x + y = 2a + 2b = 2(a+b). Therefore x+y has 2 as a factor and, by definition, is even. Hence the sum of any two even integers is even.
This proof uses the definition of even integers, the integer properties of closure under addition and multiplication, and distributivity.

Proof by mathematical induction

Despite its name, mathematical induction is a method of deduction, not a form of inductive reasoning. In proof by mathematical induction, a single "base case" is proved, and an "induction rule" is proved that establishes that any arbitrary case implies the next case. Since in principle the induction rule can be applied repeatedly starting from the proved base case, we see that all (usually infinitely many) cases are provable. This avoids having to prove each case individually. A variant of mathematical induction is proof by infinite descent, which can be used, for example, to prove the irrationality of the square root of two.

A common application of proof by mathematical induction is to prove that a property known to hold for one number holds for all natural numbers: Let N = {1,2,3,4,...} be the set of natural numbers, and P(n) be a mathematical statement involving the natural number n belonging to N such that
  • (i) P(1) is true, i.e., P(n) is true for n = 1.
  • (ii) P(n+1) is true whenever P(n) is true, i.e., P(n) is true implies that P(n+1) is true.
  • Then P(n) is true for all natural numbers n.
For example, we can prove by induction that all positive integers of the form 2n − 1 are odd. Let P(n) represent "2n − 1 is odd":
(i) For n = 1, 2n − 1 = 2(1) − 1 = 1, and 1 is odd, since it leaves a remainder of 1 when divided by 2. Thus P(1) is true.
(ii) For any n, if 2n − 1 is odd (P(n)), then (2n − 1) + 2 must also be odd, because adding 2 to an odd number results in an odd number. But (2n − 1) + 2 = 2n + 1 = 2(n+1) − 1, so 2(n+1) − 1 is odd (P(n+1)). So P(n) implies P(n+1).
Thus 2n − 1 is odd, for all positive integers n.
The shorter phrase "proof by induction" is often used instead of "proof by mathematical induction".

Proof by contraposition

Proof by contraposition infers the conclusion "if p then q" from the premise "if not q then not p". The statement "if not q then not p" is called the contrapositive of the statement "if p then q". For example, contraposition can be used to establish that, given an integer x, if x^{2} is even, then x is even:
Suppose x is not even. Then x is odd. The product of two odd numbers is odd, hence {\displaystyle x^{2}=x\cdot x} is odd. Thus x^{2} is not even. Thus, if x^{2} is even, the supposition must be false, so x has to be even.

Proof by contradiction

In proof by contradiction (also known as reductio ad absurdum, Latin for "by reduction to the absurd"), it is shown that if some statement were true, a logical contradiction occurs, hence the statement must be false. A famous example of proof by contradiction shows that {\sqrt {2}} is an irrational number:
Suppose that {\sqrt {2}} were a rational number, so by definition {\sqrt {2}}={a \over b} where a and b are non-zero integers with no common factor. (If there is a common factor, divide both numerator and denominator by that factor to remove it, and repeat until no common factor remains. By the method of infinite descent, this process must terminate.) Thus, b{\sqrt {2}}=a. Squaring both sides yields 2b2 = a2. Since 2 divides the left hand side, 2 must also divide the right hand side (otherwise an even number would equal an odd number). So a2 is even, which implies that a must also be even. So we can write a = 2c, where c is also an integer. Substitution into the original equation yields 2b2 = (2c)2 = 4c2. Dividing both sides by 2 yields b2 = 2c2. But then, by the same argument as before, 2 divides b2, so b must be even. However, if a and b are both even, they have a common factor, namely 2. This contradicts our initial supposition, so we are forced to conclude that {\sqrt {2}} is an irrational number.

Proof by construction

Proof by construction, or proof by example, is the construction of a concrete example with a property to show that something having that property exists. Joseph Liouville, for instance, proved the existence of transcendental numbers by constructing an explicit example. It can also be used to construct a counterexample to disprove a proposition that all elements have a certain property.

Proof by exhaustion

In proof by exhaustion, the conclusion is established by dividing it into a finite number of cases and proving each one separately. The number of cases sometimes can become very large. For example, the first proof of the four color theorem was a proof by exhaustion with 1,936 cases. This proof was controversial because the majority of the cases were checked by a computer program, not by hand. The shortest known proof of the four color theorem as of 2011 still has over 600 cases.

Probabilistic proof

A probabilistic proof is one in which an example is shown to exist, with certainty, by using methods of probability theory. Probabilistic proof, like proof by construction, is one of many ways to show existence theorems.

This is not to be confused with an argument that a theorem is 'probably' true, a 'plausibility argument'. The work on the Collatz conjecture shows how far plausibility is from genuine proof.

While most mathematicians do not think that probabilistic evidence ever counts as a genuine mathematical proof, a few mathematicians and philosophers have argued that at least some types of probabilistic evidence (such as Rabin's probabilistic algorithm for testing primality) are as good as genuine mathematical proofs.

Combinatorial proof

A combinatorial proof establishes the equivalence of different expressions by showing that they count the same object in different ways. Often a bijection between two sets is used to show that the expressions for their two sizes are equal. Alternatively, a double counting argument provides two different expressions for the size of a single set, again showing that the two expressions are equal.

Nonconstructive proof

A nonconstructive proof establishes that a mathematical object with a certain property exists without explaining how such an object can be found. Often, this takes the form of a proof by contradiction in which the nonexistence of the object is proved to be impossible. In contrast, a constructive proof establishes that a particular object exists by providing a method of finding it. A famous example of a nonconstructive proof shows that there exist two irrational numbers a and b such that a^{b} is a rational number:
Either {\sqrt {2}}^{\sqrt {2}} is a rational number and we are done (take a=b={\sqrt {2}}), or {\sqrt {2}}^{\sqrt {2}} is irrational so we can write a={\sqrt {2}}^{\sqrt {2}} and b={\sqrt {2}}. This then gives \left({\sqrt {2}}^{\sqrt {2}}\right)^{\sqrt {2}}={\sqrt {2}}^{2}=2, which is thus a rational of the form a^{b}.

Statistical proofs in pure mathematics

The expression "statistical proof" may be used technically or colloquially in areas of pure mathematics, such as involving cryptography, chaotic series, and probabilistic or analytic number theory. It is less commonly used to refer to a mathematical proof in the branch of mathematics known as mathematical statistics. See also "Statistical proof using data" section below.

Computer-assisted proofs

Until the twentieth century it was assumed that any proof could, in principle, be checked by a competent mathematician to confirm its validity. However, computers are now used both to prove theorems and to carry out calculations that are too long for any human or team of humans to check; the first proof of the four color theorem is an example of a computer-assisted proof. Some mathematicians are concerned that the possibility of an error in a computer program or a run-time error in its calculations calls the validity of such computer-assisted proofs into question. In practice, the chances of an error invalidating a computer-assisted proof can be reduced by incorporating redundancy and self-checks into calculations, and by developing multiple independent approaches and programs. Errors can never be completely ruled out in case of verification of a proof by humans either, especially if the proof contains natural language and requires deep mathematical insight.

Undecidable statements

A statement that is neither provable nor disprovable from a set of axioms is called undecidable (from those axioms). One example is the parallel postulate, which is neither provable nor refutable from the remaining axioms of Euclidean geometry.

Mathematicians have shown there are many statements that are neither provable nor disprovable in Zermelo-Fraenkel set theory with the axiom of choice (ZFC), the standard system of set theory in mathematics (assuming that ZFC is consistent); see list of statements undecidable in ZFC.

Gödel's (first) incompleteness theorem shows that many axiom systems of mathematical interest will have undecidable statements.

Heuristic mathematics and experimental mathematics

While early mathematicians such as Eudoxus of Cnidus did not use proofs, from Euclid to the foundational mathematics developments of the late 19th and 20th centuries, proofs were an essential part of mathematics. With the increase in computing power in the 1960s, significant work began to be done investigating mathematical objects outside of the proof-theorem framework, in experimental mathematics. Early pioneers of these methods intended the work ultimately to be embedded in a classical proof-theorem framework, e.g. the early development of fractal geometry, which was ultimately so embedded.

Related concepts

Visual proof

Although not a formal proof, a visual demonstration of a mathematical theorem is sometimes called a "proof without words". The left-hand picture below is an example of a historic visual proof of the Pythagorean theorem in the case of the (3,4,5) triangle.
Some illusory visual proofs, such as the missing square puzzle, can be constructed in a way which appear to prove a supposed mathematical fact but only do so under the presence of tiny errors (for example, supposedly straight lines which actually bend slightly) which are unnoticeable until the entire picture is closely examined, with lengths and angles precisely measured or calculated.

Elementary proof

An elementary proof is a proof which only uses basic techniques. More specifically, the term is used in number theory to refer to proofs that make no use of complex analysis. For some time it was thought that certain theorems, like the prime number theorem, could only be proved using "higher" mathematics. However, over time, many of these results have been reproved using only elementary techniques.

Two-column proof

A two-column proof published in 1913

A particular way of organising a proof using two parallel columns is often used in elementary geometry classes in the United States. The proof is written as a series of lines in two columns. In each line, the left-hand column contains a proposition, while the right-hand column contains a brief explanation of how the corresponding proposition in the left-hand column is either an axiom, a hypothesis, or can be logically derived from previous propositions. The left-hand column is typically headed "Statements" and the right-hand column is typically headed "Reasons".

Colloquial use of "mathematical proof"

The expression "mathematical proof" is used by lay people to refer to using mathematical methods or arguing with mathematical objects, such as numbers, to demonstrate something about everyday life, or when data used in an argument is numerical. It is sometimes also used to mean a "statistical proof" (below), especially when used to argue from data.

Statistical proof using data

"Statistical proof" from data refers to the application of statistics, data analysis, or Bayesian analysis to infer propositions regarding the probability of data. While using mathematical proof to establish theorems in statistics, it is usually not a mathematical proof in that the assumptions from which probability statements are derived require empirical evidence from outside mathematics to verify. In physics, in addition to statistical methods, "statistical proof" can refer to the specialized mathematical methods of physics applied to analyze data in a particle physics experiment or observational study in physical cosmology. "Statistical proof" may also refer to raw data or a convincing diagram involving data, such as scatter plots, when the data or diagram is adequately convincing without further analysis.

Inductive logic proofs and Bayesian analysis

Proofs using inductive logic, while considered mathematical in nature, seek to establish propositions with a degree of certainty, which acts in a similar manner to probability, and may be less than full certainty. Inductive logic should not be confused with mathematical induction.

Bayesian analysis uses Bayes' theorem to update a person's assessment of likelihoods of hypotheses when new evidence or information is acquired.

Proofs as mental objects

Psychologism views mathematical proofs as psychological or mental objects. Mathematician philosophers, such as Leibniz, Frege, and Carnap have variously criticized this view and attempted to develop a semantics for what they considered to be the language of thought, whereby standards of mathematical proof might be applied to empirical science.

Influence of mathematical proof methods outside mathematics

Philosopher-mathematicians such as Spinoza have attempted to formulate philosophical arguments in an axiomatic manner, whereby mathematical proof standards could be applied to argumentation in general philosophy. Other mathematician-philosophers have tried to use standards of mathematical proof and reason, without empiricism, to arrive at statements outside of mathematics, but having the certainty of propositions deduced in a mathematical proof, such as Descartes' cogito argument.

Ending a proof

Sometimes, the abbreviation "Q.E.D." is written to indicate the end of a proof. This abbreviation stands for "Quod Erat Demonstrandum", which is Latin for "that which was to be demonstrated". A more common[citation needed] alternative is to use a square or a rectangle, such as □ or ∎, known as a "tombstone" or "halmos" after its eponym Paul Halmos. Often, "which was to be shown" is verbally stated when writing "QED", "□", or "∎" during an oral presentation.

MIT Assesses the Technical Feasibility of the Mars One Mission

By
Assessing the Technical Feasibility of the Proposed Mars One Mission
The non-profit company Mars One plans to establish the first human settlement on Mars by 2025. Pictured is an artist’s rendering of a series of habitats. Solar panels (in the foreground), would supply the colony’s electricity, while a system to extract water from the soil (in the background) would supply drinking water.

A new independent assessment of the Mars One Mission from MIT engineers reveals that the project may have to take a step back, at least to reconsider the mission’s technical feasibility.

In 2012, the “Mars One” project, led by a Dutch nonprofit, announced plans to establish the first human colony on the Red Planet by 2025. The mission would initially send four astronauts on a one-way trip to Mars, where they would spend the rest of their lives building the first permanent human settlement.

It’s a bold vision — particularly since Mars One claims that the entire mission can be built upon technologies that already exist. As its website states, establishing humans on Mars would be “the next giant leap for mankind.”

But engineers at MIT say the project may have to take a step back, at least to reconsider the mission’s technical feasibility.

The MIT researchers developed a detailed settlement-analysis tool to assess the feasibility of the Mars One mission, and found that new technologies will be needed to keep humans alive on Mars.

For example, if all food is obtained from locally grown crops, as Mars One envisions, the vegetation would produce unsafe levels of oxygen, which would set off a series of events that would eventually cause human inhabitants to suffocate. To avoid this scenario, a system to remove excess oxygen would have to be implemented — a technology that has not yet been developed for use in space.

Similarly, the Mars Phoenix lander discovered evidence of ice on the Martian surface in 2008, suggesting that future settlers might be able to melt ice for drinking water — another Mars One goal. But according to the MIT analysis, current technologies designed to “bake” water from soil are not yet ready for deployment, particularly in space.

The team also performed an integrated analysis of spare-parts resupply — how many spare parts would have to be delivered to a Martian colony at each opportunity to keep it going. The researchers found that as the colony grows, spare parts would quickly dominate future deliveries to Mars, making up as much as 62 percent of payloads from Earth.

As for the actual voyage to Mars, the team also calculated the number of rockets required to establish the first four settlers and subsequent crews on the planet, as well as the journey’s cost.

According to the Mars One plan, six Falcon Heavy rockets would be required to send up initial supplies, before the astronauts’ arrival. But the MIT assessment found that number to be “overly optimistic”: The team determined that the needed supplies would instead require 15 Falcon Heavy rockets. The transportation cost for this leg of the mission alone, combined with the astronauts’ launch, would be $4.5 billion — a cost that would grow with additional crews and supplies to Mars. The researchers say this estimate does not include the cost of developing and purchasing equipment for the mission, which would further increase the overall cost.

Olivier de Weck, an MIT professor of aeronautics and astronautics and engineering systems, says the prospect of building a human settlement on Mars is an exciting one. To make this goal a reality, however, will require innovations in a number of technologies and a rigorous systems perspective, he says.

“We’re not saying, black and white, Mars One is infeasible,” de Weck says. “But we do think it’s not really feasible under the assumptions they’ve made. We’re pointing to technologies that could be helpful to invest in with high priority, to move them along the feasibility path.”

“One of the great insights we were able to get was just how hard it is to pull this [mission] off,” says graduate student Sydney Do. “There are just so many unknowns. And to give anyone confidence that they’re going to get there and stay alive — there’s still a lot of work that needs to be done.”

Do and de Weck presented their analysis this month at the International Astronautical Congress in Toronto. Co-authors include MIT graduate students Koki Ho, Andrew Owens, and Samuel Schreiner.

Simulating a day on Mars

The group took a systems-based approach in analyzing the Mars One mission, first assessing various aspects of the mission’s architecture, such as its habitat, life-support systems, spare-parts requirements, and transportation logistics, then looking at how each component contributes to the whole system.

For the habitat portion, Do simulated the day-to-day life of a Mars colonist. Based on the typical work schedule, activity levels, and metabolic rates of astronauts on the International Space Station (ISS), Do estimated that a settler would have to consume about 3,040 calories daily to stay alive and healthy on Mars. He then determined crops that would provide a reasonably balanced diet, including beans, lettuce, peanuts, potatoes, and rice.

Do calculated that producing enough of these crops to sustain astronauts over the long term would require about 200 square meters of growing area, compared with Mars One’s estimate of 50 square meters. If, as the project plans, crops are cultivated within the settlers’ habitat, Do found that they would produce unsafe levels of oxygen that would exceed fire safety thresholds, requiring continuous introduction of nitrogen to reduce the oxygen level. Over time, this would deplete nitrogen tanks, leaving the habitat without a gas to compensate for leaks.

As the air inside the habitat continued to leak, the total atmospheric pressure would drop, creating an oppressive environment that would suffocate the first settler within an estimated 68 days.

Possible solutions, Do says, might include either developing a technology to extract excess oxygen or isolating the crops in a separate greenhouse. The team even considered using nitrogen extracted from the Martian atmosphere, but found that doing so would require a prohibitively large system. Surprisingly, the cheapest option found was to supply all the food required from Earth.

“We found carrying food is always cheaper than growing it locally,” Do says. “On Mars, you need lighting and watering systems, and for lighting, we found it requires 875 LED systems, which fail over time. So you need to provide spare parts for that, making the initial system heavier.”

Twisting the knobs

As the team found, spare parts, over time, would substantially inflate the cost of initial and future missions to Mars. Owens, who assessed the resupply of spare parts, based his analysis on reliability data derived from NASA repair logs for given components on the ISS.

“The ISS is based on the idea that if something breaks, you can call home and get a new part quickly,” says Owens. “If you want a spare part on Mars, you have to send it when a launch window is open, every 26 months, and then wait 180 days for it to get there. If you could make spares in-situ, that would be a massive savings.”

Owens points to technologies such as 3-D printing, which may enable settlers to manufacture spare parts on Mars. But the technology as it exists today is not advanced enough to reproduce the exact dimensions and functions of many space-rated parts. The MIT analysis found that 3-D printers will have to improve by leaps, or else the entire Mars settlement infrastructure will have to be redesigned so that its parts can be printed with existing technology.

While this analysis may make the Mars One program look daunting, the researchers say the settlement-analysis tool they’ve developed may help determine the feasibility of various scenarios. For example, rather than sending crews on one-way trips to the planet, what would the overall mission cost be if crews were occasionally replaced?

“Mars One is a pretty radical idea,” Schreiner says. “Now we’ve built a tool that we can play around with, and we can twist some of the knobs to see how the cost and feasibility of the mission changes.”

Tracy Gill, a technology strategy manager at NASA, says the tool may be applicable for assessing other missions to Mars, and points to a few scenarios that the group may want to explore using the settlement-analysis tool.

“This [tool] can provide a benefit to mission planners by allowing them to evaluate a larger spectrum of mission architectures with better confidence in their analysis,” says Gill, who did not contribute to the research. “Included among those architectures would be options ranging from completely growing all food in situ with bioregenerative systems, to packaging all food products from Earth, to various combination of those two extremes.“

Some of the students on this project were supported by NASA fellowships.

Related Research Paper: An Independent Assessment of the Technical Feasibility of the Mars One Mission Plan

Image: Bryan Versteeg/Mars One

Megatsunami

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Meg...