In logic and philosophy, an argument is a series of statements (in a natural language), called the premises, intended to determine the degree of truth of another statement, the conclusion. The logical form of an argument in a natural language can be represented in a symbolic formal language, and independently of natural language formally defined "arguments" can be made in math and computer science.
Logic is the study of the forms of reasoning in arguments and the development of standards and criteria to evaluate arguments. Deductive arguments can be valid or sound:
in a valid argument, premises necessitate the conclusion, even if one
or more of the premises is false and the conclusion is false; in a sound
argument, true premises necessitate a true conclusion. Inductive arguments,
by contrast, can have different degrees of logical strength: the
stronger or more cogent the argument, the greater the probability that
the conclusion is true, the weaker the argument, the lesser that
probability.
The standards for evaluating non-deductive arguments may rest on
different or additional criteria than truth—for example, the
persuasiveness of so-called "indispensability claims" in transcendental arguments, the quality of hypotheses in retroduction, or even the disclosure of new possibilities for thinking and acting.
Etymology
The Latin root arguere (to make bright, enlighten, make known, prove, etc.) is from Proto-Indo-Europeanargu-yo-, suffixed form of arg- (to shine; white).
Formal and informal
Informal arguments as studied in informal logic, are presented in ordinary language and are intended for everyday discourse. Formal arguments are studied in formal logic (historically called symbolic logic, more commonly referred to as mathematical logic today) and are expressed in a formal language. Informal logic emphasizes the study of argumentation; formal logic emphasizes implication and inference.
Informal arguments are sometimes implicit. The rational structure –
the relationship of claims, premises, warrants, relations of
implication, and conclusion – is not always spelled out and immediately
visible and must be made explicit by analysis.
Standard types
Argument terminology
There are several kinds of arguments in logic, the best-known of
which are "deductive" and "inductive." An argument has one or more
premises but only one conclusion. Each premise and the conclusion are truth bearers
or "truth-candidates", each capable of being either true or false (but
not both). These truth values bear on the terminology used with
arguments.
Deductive arguments
A deductive argument asserts that the truth of the conclusion is a logical consequence
of the premises. Based on the premises, the conclusion follows
necessarily (with certainty). For example, given premises that A=B and
B=C, then the conclusion follows necessarily that A=C. Deductive
arguments are sometimes referred to as "truth-preserving" arguments.
A deductive argument is said to be valid
or invalid. If one assumes the premises to be true (ignoring their
actual truth values), would the conclusion follow with certainty? If
yes, the argument is valid. If no, it is invalid. In determining
validity, the structure of the argument is essential to the
determination, not the actual truth values. For example, consider the
argument that because bats can fly (premise=true), and all flying
creatures are birds (premise=false), therefore bats are birds
(conclusion=false). If we assume the premises are true, the conclusion
follows necessarily, and it is a valid argument.
If a deductive argument is valid and its premises are all true, then
it is also referred to as sound. Otherwise, it is unsound, as "bats are
birds".
If all the premises of a valid deductive argument are true, then its
conclusion must be true. It is impossible for the conclusion to be
false if all the premises are true.
Inductive arguments
An inductive argument
asserts that the truth of the conclusion is supported by the
probability of the premises. For example, given that the U.S. military
budget is the largest in the world (premise=true), then it is probable
that it will remain so for the next 10 years (conclusion=true).
Arguments that involve predictions are inductive since the future is
uncertain.
An inductive argument is said to be strong or weak. If the premises
of an inductive argument are assumed true, is it probable the conclusion
is also true? If yes, the argument is strong. If no, it is weak.
A strong argument is said to be cogent if it has all true premises.
Otherwise, the argument is uncogent. The military budget argument
example is a strong, cogent argument.
Deductive
A deductive argument, if valid, has a conclusion that is entailed
by its premises. The truth of the conclusion is a logical consequence
of the premises If the premises are true, the conclusion must be true.
It would be self-contradictory to assert the premises and deny the
conclusion, because negation of the conclusion is contradictory to the
truth of the premises.
Validity
Deductive arguments may be either valid or invalid. If an argument is
valid, it is a valid deduction, and if its premises are true, the
conclusion must be true: a valid argument cannot have true premises and a
false conclusion.
An argument is formally valid if and only if the denial of the conclusion is incompatible with accepting all the premises.
The validity of an argument depends not on the actual truth or
falsity of its premises and conclusion, but on whether the argument has a
valid logical form.
The validity of an argument is not a guarantee of the truth of its
conclusion. A valid argument may have false premises that render it
inconclusive: the conclusion of a valid argument with one or more false
premises may be true or false.
Logic seeks to discover the forms that make arguments valid. A
form of argument is valid if and only if the conclusion is true under
all interpretations of that argument in which the premises are true.
Since the validity of an argument depends on its form, an argument can
be shown invalid by showing that its form is invalid. This can be done
by a counter example of the same form of argument with premises that are
true under a given interpretation, but a conclusion that is false under
that interpretation. In informal logic this is called a counter argument.
The form of argument can be shown by the use of symbols. For each
argument form, there is a corresponding statement form, called a corresponding conditional, and an argument form is valid if and only if its corresponding conditional is a logical truth.
A statement form which is logically true is also said to be a valid
statement form. A statement form is a logical truth if it is true under
all interpretations. A statement form can be shown to be a logical truth by either (a) showing that it is a tautology or (b) by means of a proof procedure.
The corresponding conditional of a valid argument is a necessary truth (true in all possible worlds)
and so the conclusion necessarily follows from the premises, or follows
of logical necessity. The conclusion of a valid argument is not
necessarily true, it depends on whether the premises are true. If the
conclusion, itself, is a necessary truth, it is without regard to the
premises.
Some examples:
All Greeks are human and all humans are mortal; therefore, all Greeks are mortal. : Valid argument; if the premises are true the conclusion must be true.
Some Greeks are logicians and some logicians are tiresome; therefore, some Greeks are tiresome. Invalid argument: the tiresome logicians might all be Romans (for example).
Either we are all doomed or we are all saved; we are not all saved; therefore, we are all doomed.
Valid argument; the premises entail the conclusion. (This does not
mean the conclusion has to be true; it is only true if the premises are
true, which they may not be!)
Some men are hawkers. Some hawkers are rich. Therefore, some men are rich. Invalid argument. This can be easier seen by giving a counter-example with the same argument form:
Some people are herbivores. Some herbivores are zebras. Therefore, some people are zebras. Invalid argument, as it is possible that the premises be true and the conclusion false.
In the above second to last case (Some men are hawkers...), the
counter-example follows the same logical form as the previous argument,
(Premise 1: "Some X are Y." Premise 2: "Some Y are Z." Conclusion: "Some X are Z.")
in order to demonstrate that whatever hawkers may be, they may or may
not be rich, in consideration of the premises as such.
The forms of argument that render deductions valid are
well-established, however some invalid arguments can also be persuasive
depending on their construction (inductive arguments, for example).
Soundness
A sound argument is a valid argument whose conclusion follows from its premise(s), and the premise(s) of which is/are true.
Inductive
Non-deductive logic is reasoning using arguments in which the
premises support the conclusion but do not entail it. Forms of
non-deductive logic include the statistical syllogism, which argues from generalizations true for the most part, and induction, a form of reasoning that makes generalizations based on individual instances. An inductive argument is said to be cogent if and only if the truth of the argument's premises would render the truth of the conclusion probable (i.e., the argument is strong), and the argument's premises are, in fact, true. Cogency can be considered inductive logic's analogue to deductive logic's "soundness". Despite its name, mathematical induction is not a form of inductive reasoning. The lack of deductive validity is known as the problem of induction.
Defeasible arguments and argumentation schemes
In modern argumentation theories, arguments are regarded as defeasible passages from premises to a conclusion. Defeasibility
means that when additional information (new evidence or contrary
arguments) is provided, the premises may be no longer lead to the
conclusion (non-monotonic reasoning). This type of reasoning is referred to as defeasible reasoning. For instance we consider the famous Tweety example:
Tweety is a bird.
Birds generally fly.
Therefore, Tweety (probably) flies.
This argument is reasonable and the premises support the conclusion
unless additional information indicating that the case is an exception
comes in. If Tweety is a penguin, the inference is no longer justified
by the premise. Defeasible arguments are based on generalizations that
hold only in the majority of cases, but are subject to exceptions and
defaults.
In order to represent and assess defeasible reasoning, it is
necessary to combine the logical rules (governing the acceptance of a
conclusion based on the acceptance of its premises) with rules of
material inference, governing how a premise can support a given
conclusion (whether it is reasonable or not to draw a specific
conclusion from a specific description of a state of affairs).
Argumentation schemes have been developed to describe and assess the acceptability or the fallaciousness
of defeasible arguments. Argumentation schemes are stereotypical
patterns of inference, combining semantic-ontological relations with
types of reasoning and logical axioms and representing the abstract
structure of the most common types of natural arguments. A typical example is the argument from expert opinion, shown below, which has two premises and a conclusion.
Argument from expert opinion
Major Premise:
Source E is an expert in subject domain S containing proposition A.
Minor Premise:
E asserts that proposition A is true (false).
Conclusion:
A is true (false).
Each scheme may be associated with a set of critical questions,
namely criteria for assessing dialectically the reasonableness and
acceptability of an argument. The matching critical questions are the
standard ways of casting the argument into doubt.
By analogy
Argument by analogy
may be thought of as argument from the particular to particular. An
argument by analogy may use a particular truth in a premise to argue
towards a similar particular truth in the conclusion. For example, if A.
Plato was mortal, and B. Socrates was like Plato in other respects,
then asserting that C. Socrates was mortal is an example of argument by
analogy because the reasoning employed in it proceeds from a particular
truth in a premise (Plato was mortal) to a similar particular truth in
the conclusion, namely that Socrates was mortal.
Other kinds
Other kinds of arguments may have different or additional standards of validity or justification. For example, philosopher Charles Taylor said that so-called transcendental arguments
are made up of a "chain of indispensability claims" that attempt to
show why something is necessarily true based on its connection to our
experience, while Nikolas Kompridis has suggested that there are two types of "fallible" arguments: one based on truth claims, and the other based on the time-responsive disclosure of possibility (world disclosure). Kompridis said that the French philosopher Michel Foucault was a prominent advocate of this latter form of philosophical argument.
World-disclosing
World-disclosing arguments are a group of philosophical arguments that according to Nikolas Kompridis employ a disclosive approach, to reveal features of a wider ontological
or cultural-linguistic understanding – a "world", in a specifically
ontological sense – in order to clarify or transform the background of
meaning (tacit knowledge) and what Kompridis has called the "logical space" on which an argument implicitly depends.
Explanations
While arguments attempt to show that something was, is, will be, or should be the case, explanations try to show why or how something is or will be. If Fred and Joe address the issue of whether
or not Fred's cat has fleas, Joe may state: "Fred, your cat has fleas.
Observe, the cat is scratching right now." Joe has made an argument that
the cat has fleas. However, if Joe asks Fred, "Why is your cat
scratching itself?" the explanation, "...because it has fleas." provides
understanding.
Both the above argument and explanation require knowing the
generalities that a) fleas often cause itching, and b) that one often
scratches to relieve itching. The difference is in the intent: an
argument attempts to settle whether or not some claim
is true, and an explanation attempts to provide understanding of the
event. Note, that by subsuming the specific event (of Fred's cat
scratching) as an instance of the general rule that "animals scratch
themselves when they have fleas", Joe will no longer wonder why
Fred's cat is scratching itself. Arguments address problems of belief,
explanations address problems of understanding. Also note that in the
argument above, the statement, "Fred's cat has fleas" is up for debate
(i.e. is a claim), but in the explanation, the statement, "Fred's cat
has fleas" is assumed to be true (unquestioned at this time) and just
needs explaining.
Arguments and explanations largely resemble each other in rhetorical use. This is the cause of much difficulty in thinking critically about claims. There are several reasons for this difficulty.
People often are not themselves clear on whether they are arguing for or explaining something.
The same types of words and phrases are used in presenting explanations and arguments.
The terms 'explain' or 'explanation,' et cetera are frequently used in arguments.
Explanations are often used within arguments and presented so as to serve as arguments.
Likewise, "...arguments are essential to the process of justifying
the validity of any explanation as there are often multiple explanations
for any given phenomenon."
Explanations and arguments are often studied in the field of Information Systems
to help explain user acceptance of knowledge-based systems. Certain
argument types may fit better with personality traits to enhance
acceptance by individuals.
Fallacies and non-arguments
Fallacies are types of argument or expressions which are held to be of an invalid form or contain errors in reasoning.
One type of fallacy occurs when a word frequently used to
indicate a conclusion is used as a transition (conjunctive adverb)
between independent clauses. In English the words therefore, so, because and hence typically separate the premises from the conclusion of an argument. Thus: Socrates is a man, all men are mortal therefore Socrates is mortal is an argument because the assertion Socrates is mortal follows from the preceding statements. However, I was thirsty and therefore I drank is not an argument, despite its appearance. It is not being claimed that I drank is logically entailed by I was thirsty. The therefore in this sentence indicates for that reason not it follows that.
Elliptical or ethymematic arguments
Often
an argument is invalid or weak because there is a missing premise—the
supply of which would make it valid or strong. This is referred to as an
elliptical or ethymematic argument.
Speakers and writers will often leave out a necessary premise in their
reasoning if it is widely accepted and the writer does not wish to state
the blindingly obvious. Example: All metals expand when heated, therefore iron will expand when heated. The missing premise is: Iron is a metal.
On the other hand, a seemingly valid argument may be found to lack a
premise – a "hidden assumption" – which, if highlighted, can show a
fault in reasoning. Example: A witness reasoned: Nobody came out the front door except the milkman; therefore the murderer must have left by the back door. The hidden assumptions are: (1) the milkman was not the murderer and (2) the murderer has left by the front or back door.
Argument mining
The goal of argument mining is the automatic extraction and identification of argumentative structures from natural language text with the aid of computer programs. Such argumentative structures include the premise, conclusions, the argument scheme and the relationship between the main and subsidiary argument, or the main and counter-argument within discourse.
Logic (from Greek: λογική, logikḗ, 'possessed of reason, intellectual, dialectical, argumentative') is the systematic study of valid rules of inference, i.e. the relations that lead to the acceptance of one proposition (the conclusion) on the basis of a set of other propositions (premises). More broadly, logic is the analysis and appraisal of arguments.There is no universal agreement as to the exact definition and boundaries of logic.
A good argument not only possesses validity and soundness (or strength, in induction), but it also avoids circular dependencies, is clearly stated, relevant, and consistent; otherwise it is useless for reasoning and persuasion, and is classified as a fallacy.
In ordinary discourse, inferences may be signified by words such as therefore, thus, hence, ergo, and so on.
Historically, logic has been studied in philosophy (since ancient times) and mathematics (since the mid-19th century). More recently, logic has been studied in cognitive science, which draws on computer science, linguistics, philosophy and psychology, among other disciplines. A logician is any person, often a philosopher or mathematician, whose topic of scholarly study is logic.
Types of logic
Upon this first, and
in one sense this sole, rule of reason, that in order to learn you must
desire to learn, and in so desiring not be satisfied with what you
already incline to capably think, there follows one corollary which
itself deserves to be inscribed upon every wall of the city of
philosophy: Do not block the way of inquiry.
Philosophical logic
is an area of philosophy. It's a set of methods used to solve
philosophical problems and a fundamental tool for the advancement of metaphilosophy.
Informal logic
Informal logic is the study of natural languagearguments. The study of fallacies
is an important branch of informal logic. Since much informal argument
is not strictly speaking deductive, on some conceptions of logic,
informal logic is not logic at all.
Formal logic
Formal logic is the study of inference with purely formal content. An inference possesses a purely formal and explicit content
(i.e. it can be expressed as a particular application of a wholly
abstract rule) such as, a rule that is not about any particular thing or
property. In many definitions of logic, logical consequence and inference with purely formal content are the same.
Syllogistic logic can be found in the works of Aristotle, making it the earliest known formal study and studies types of syllogism. Modern formal logic follows and expands on Aristotle.
Symbolic logic is the study of symbolic abstractions that capture the formal features of logical inference, often divided into two main branches: propositional logic and predicate logic.
An argument is constructed by applying one of the forms of the different types of logical reasoning: deductive, inductive, and abductive. In deduction, the validity of an argument is determined solely by its logical form, not its content, whereas the soundness requires both validity and that all the given premises are actually true.
Completeness, consistency, decidability, and expressivity, are further fundamental concepts in logic. The categorization of the logical systems and of their properties has led to the emergence of a metatheory of logic known as metalogic. However, agreement on what logic actually is has remained elusive, although the field of universal logic has studied the common structure of logics.
Logical form
Logic is generally considered formal when it analyzes and represents the form of any valid argument
type. The form of an argument is displayed by representing its
sentences in the formal grammar and symbolism of a logical language to
make its content usable in formal inference. Simply put, to formalize
simply means to translate English sentences into the language of logic.
This is called showing the logical form of the argument.
It is necessary because indicative sentences of ordinary language show a
considerable variety of form and complexity that makes their use in
inference impractical. It requires, first, ignoring those grammatical
features irrelevant to logic (such as gender and declension, if the
argument is in Latin), replacing conjunctions irrelevant to logic (e.g.
"but") with logical conjunctions
like "and" and replacing ambiguous, or alternative logical expressions
("any", "every", etc.) with expressions of a standard type (e.g. "all",
or the universal quantifier ∀).
Second, certain parts of the sentence must be replaced with
schematic letters. Thus, for example, the expression "all Ps are Qs"
shows the logical form common to the sentences "all men are mortals",
"all cats are carnivores", "all Greeks are philosophers", and so on.
The schema can further be condensed into the formulaA(P,Q), where the letter A indicates the judgement 'all – are –'.
The importance of form was recognised from ancient times. Aristotle uses variable letters to represent valid inferences in Prior Analytics, leading Jan Łukasiewicz to say that the introduction of variables was "one of Aristotle's greatest inventions". According to the followers of Aristotle (such as Ammonius),
only the logical principles stated in schematic terms belong to logic,
not those given in concrete terms. The concrete terms 'man', 'mortal',
etc., are analogous to the substitution values of the schematic
placeholders P, Q, R, which were called the 'matter' (Greek: ὕλη, hyle) of the inference.
There is a big difference between the kinds of formulas seen in traditional term logic and the predicate calculus that is the fundamental advance of modern logic. The formula A(P,Q) (all Ps are Qs) of traditional logic corresponds to the more complex formula in predicate logic, involving the logical connectives for universal quantification and implication rather than just the predicate letter A and using variable arguments where traditional logic uses just the term letter P. With the complexity comes power, and the advent of the predicate calculus inaugurated revolutionary growth of the subject.
Semantics
The validity of an argument depends upon the meaning, or semantics, of the sentences that make it up.
Aristotle's six Organon, especially De Interpretatione, gives a cursory outline of semantics which the scholastic logicians, particularly in the thirteenth and fourteenth century, developed into a complex and sophisticated theory, called supposition theory.
This showed how the truth of simple sentences, expressed schematically,
depend on how the terms 'supposit', or stand for, certain
extra-linguistic items. For example, in part II of his Summa Logicae, William of Ockham presents a comprehensive account of the necessary and sufficient conditions for the truth
of simple sentences, in order to show which arguments are valid and
which are not. Thus "every A is B' is true if and only if there is
something for which 'A' stands, and there is nothing for which 'A'
stands, for which 'B' does not also stand."
Early modern logic defined semantics purely as a relation between ideas. Antoine Arnauld in the Port Royal-Logic,
says that after conceiving things by our ideas, we compare these ideas,
and, finding that some belong together and some do not, we unite or
separate them. This is called affirming or denying, and in general judging. Thus truth and falsity are no more than the agreement or disagreement of ideas. This suggests obvious difficulties, leading Locke
to distinguish between 'real' truth, when our ideas have 'real
existence' and 'imaginary' or 'verbal' truth, where ideas like harpies
or centaurs exist only in the mind. This view, known as psychologism,
was taken to the extreme in the nineteenth century, and is generally
held by modern logicians to signify a low point in the decline of logic
before the twentieth century.
Modern semantics is in some ways closer to the medieval view, in rejecting such psychological truth-conditions. However, the introduction of quantification, needed to solve the problem of multiple generality, rendered impossible the kind of subject-predicate analysis that underlies medieval semantics. The main modern approach is model-theoretic semantics, based on Alfred Tarski's semantic theory of truth.
The approach assumes that the meaning of the various parts of the
propositions are given by the possible ways we can give a recursively
specified group of interpretation functions from them to some predefined domain of discourse: an interpretation of first-order predicate logic is given by a mapping from terms to a universe of individuals,
and a mapping from propositions to the truth values "true" and "false".
Model-theoretic semantics is one of the fundamental concepts of model theory. Modern semantics also admits rival approaches, such as the proof-theoretic semantics
that associates the meaning of propositions with the roles that they
can play in inferences, an approach that ultimately derives from the
work of Gerhard Gentzen on structural proof theory and is heavily influenced by Ludwig Wittgenstein's later philosophy, especially his aphorism "meaning is use."
Inference
Inference is not to be confused with implication. An implication is a sentence of the form 'If p then q', and can be true or false. The stoic logicianPhilo of Megara was the first to define the truth conditions of such an implication: false only when the antecedent p is true and the consequent q is false, in all other cases true. An inference,
on the other hand, consists of two separately asserted propositions of
the form 'p therefore q'. An inference is not true or false, but valid
or invalid. However, there is a connection between implication and
inference, as follows: if the implication 'if p then q' is true, the inference 'p therefore q' is valid.
This was given an apparently paradoxical formulation by Philo, who said
that the implication 'if it is day, it is night' is true only at night,
so the inference 'it is day, therefore it is night' is valid in the
night, but not in the day.
The theory of inference (or 'consequences') was systematically developed in medieval times by logicians such as William of Ockham and Walter Burley. It is uniquely medieval, though it has its origins in Aristotle's Topica and Boethius' De Syllogismis hypotheticis.
Many terms in logic, for this reason, are in Latin. For instance, the
rule that licenses the move from the implication 'if p then q' plus the
assertion of its antecedent p, to the assertion of the consequent q, is
known as modus ponens ('mode of positing')—from Latin: posito antecedente ponitur consequens. The Latin formulations of many other rules such as ex falso quodlibet ('from falsehood, anything [follows]'), and reductio ad absurdum ('reduction to absurdity'; i.e. to disprove by showing the consequence as absurd), also date from this period.
However, the theory of consequences, or the so-called hypothetical syllogism, was never fully integrated into the theory of the categorical
syllogism. This was partly because of the resistance to reducing the
categorical judgment 'every s is p' to the so-called hypothetical
judgment 'if anything is s, it is p'. The first was thought to imply
'some s is p', the latter was not, and as late as 1911 in the Encyclopædia Britannica
article on "Logic", we find the Oxford logician T. H. Case arguing
against Sigwart's and Brentano's modern analysis of the universal
proposition.
Logical systems
A formal system is an organization
of terms used for the analysis of deduction. It consists of an
alphabet, a language over the alphabet to construct sentences, and a
rule for deriving sentences. Among the important properties that logical systems can have are:
Consistency: no theorem of the system contradicts another.
Validity: the system's rules of proof never allow a false inference from true premises.
Completeness: if a formula is true, it can be proven, i.e. is a theorem of the system.
Soundness:
if any formula is a theorem of the system, it is true. This is the
converse of completeness. (Note that in a distinct philosophical use of
the term, an argument is sound when it is both valid and its premises
are true.)
Expressivity: what concepts can be expressed in the system.
Some logical systems do not have all these properties. As an example, Kurt Gödel's incompleteness theorems show that sufficiently complex formal systems of arithmetic cannot be consistent and complete; however, first-order predicate logics not extended by specific axioms to be arithmetic formal systems with equality can be complete and consistent.
Logic and rationality
As the study of argument is of clear importance to the reasons that
we hold things to be true, logic is of essential importance to rationality.
Here we have defined logic to be "the systematic study of the form of
arguments;" the reasoning behind argument is of several sorts, but only
some of these arguments fall under the aegis of logic proper.
Deductive reasoning concerns the logical consequence
of given premises and is the form of reasoning most closely connected
to logic. On a narrow conception of logic (see below) logic concerns
just deductive reasoning, although such a narrow conception
controversially excludes most of what is called informal logic from the
discipline.
There are other forms of reasoning that are rational but that are generally not taken to be part of logic. These include inductive reasoning, which covers forms of inference that move from collections of particular judgements to universal judgements, and abductive reasoning, which is a form of inference that goes from observation to a hypothesis that accounts for the reliable data (observation) and seeks to explain relevant evidence. American philosopher Charles Sanders Peirce (1839–1914) first introduced the term as guessing. Peirce said that to abduce a hypothetical explanation from an observed surprising circumstance is to surmise that may be true because then would be a matter of course. Thus, to abduce from involves determining that is sufficient (or nearly sufficient), but not necessary, for .
While inductive and abductive inference are not part of logic
proper, the methodology of logic has been applied to them with some
degree of success. For example, the notion of deductive validity (where
an inference is deductively valid if and only if
there is no possible situation in which all the premises are true but
the conclusion false) exists in an analogy to the notion of inductive
validity, or "strength", where an inference is inductively strong if and
only if its premises give some degree of probability to its conclusion.
Whereas the notion of deductive validity can be rigorously stated for
systems of formal logic in terms of the well-understood notions of semantics,
inductive validity requires us to define a reliable generalization of
some set of observations. The task of providing this definition may be
approached in various ways, some less formal than others; some of these
definitions may use logical association rule induction, while others may use mathematical models of probability such as decision trees.
Rival conceptions
Logic arose (see below) from a concern with correctness of argumentation.
Modern logicians usually wish to ensure that logic studies just those
arguments that arise from appropriately general forms of inference. For
example, Thomas Hofweber writes in the Stanford Encyclopedia of Philosophy that logic "does not, however, cover good reasoning as a whole. That is the job of the theory of rationality.
Rather it deals with inferences whose validity can be traced back to
the formal features of the representations that are involved in that
inference, be they linguistic, mental, or other representations."
The idea that logic treats special forms of argument, deductive
argument, rather than argument in general, has a history in logic that
dates back at least to logicism
in mathematics (19th and 20th centuries) and the advent of the
influence of mathematical logic on philosophy. A consequence of taking
logic to treat special kinds of argument is that it leads to
identification of special kinds of truth, the logical truths (with logic
equivalently being the study of logical truth), and excludes many of
the original objects of study of logic that are treated as informal
logic. Robert Brandom
has argued against the idea that logic is the study of a special kind
of logical truth, arguing that instead one can talk of the logic of material inference (in the terminology of Wilfred Sellars), with logic making explicit the commitments that were originally implicit in informal inference.
Logic comes from the Greek word logos, originally meaning "the
word" or "what is spoken", but coming to mean "thought" or "reason". In
the Western World, logic was first developed by Aristotle, who called the subject 'analytics'. Aristotelian logic became widely accepted in science and mathematics and remained in wide use in the West until the early 19th century. Aristotle's system of logic was responsible for the introduction of hypothetical syllogism, temporalmodal logic, and inductive logic, as well as influential vocabulary such as terms, predicables, syllogisms and propositions. There was also the rival Stoic logic.
In Europe during the later medieval period, major efforts were made to show that Aristotle's ideas were compatible with Christian faith. During the High Middle Ages,
logic became a main focus of philosophers, who would engage in critical
logical analyses of philosophical arguments, often using variations of
the methodology of scholasticism. In 1323, William of Ockham's influential Summa Logicae
was released. By the 18th century, the structured approach to arguments
had degenerated and fallen out of favour, as depicted in Holberg's satirical play Erasmus Montanus.
The Chinese logical philosopher Gongsun Long (c. 325–250 BCE) proposed the paradox "One and one cannot become two, since neither becomes two." In China, the tradition of scholarly investigation into logic, however, was repressed by the Qin dynasty following the legalist philosophy of Han Feizi.
In India, the Anviksiki school of logic was founded by Medhātithi (c. 6th century BCE). Innovations in the scholastic school, called Nyaya, continued from ancient times into the early 18th century with the Navya-Nyāya school. By the 16th century, it developed theories resembling modern logic, such as Gottlob Frege's
"distinction between sense and reference of proper names" and his
"definition of number", as well as the theory of "restrictive conditions
for universals" anticipating some of the developments in modern set theory.
Since 1824, Indian logic attracted the attention of many Western
scholars, and has had an influence on important 19th-century logicians
such as Charles Babbage, Augustus De Morgan, and George Boole. In the 20th century, Western philosophers like Stanislaw Schayer and Klaus Glashoff have explored Indian logic more extensively.
The syllogistic logic developed by Aristotle predominated in the West until the mid-19th century, when interest in the foundations of mathematics stimulated the development of symbolic logic (now called mathematical logic). In 1854, George Boole published The Laws of Thought, introducing symbolic logic and the principles of what is now known as Boolean logic. In 1879, Gottlob Frege published Begriffsschrift, which inaugurated modern logic with the invention of quantifier
notation, reconciling the Aristotelian and Stoic logics in a broader
system, and solving such problems for which Aristotelian logic was
impotent, such as the problem of multiple generality. From 1910 to 1913, Alfred North Whitehead and Bertrand Russell published Principia Mathematica on the foundations of mathematics, attempting to derive mathematical truths from axioms and inference rules in symbolic logic. In 1931, Gödel raised serious problems with the foundationalist program and logic ceased to focus on such issues.
The development of logic since Frege, Russell, and Wittgenstein
had a profound influence on the practice of philosophy and the perceived
nature of philosophical problems and philosophy of mathematics. Logic, especially sentential logic, is implemented in computer logic circuits and is fundamental to computer science.
Logic is commonly taught by university philosophy, sociology,
advertising and literature departments, often as a compulsory
discipline.
Types
Syllogistic logic
A depiction from the 15th century of the square of opposition, which expresses the fundamental dualities of syllogistic.
The Organon was Aristotle's body of work on logic, with the Prior Analytics constituting the first explicit work in formal logic, introducing the syllogistic. The parts of syllogistic logic, also known by the name term logic,
are the analysis of the judgements into propositions consisting of two
terms that are related by one of a fixed number of relations, and the
expression of inferences by means of syllogisms
that consist of two propositions sharing a common term as premise, and a
conclusion that is a proposition involving the two unrelated terms from
the premises.
Aristotle's work was regarded in classical times and from
medieval times in Europe and the Middle East as the very picture of a
fully worked out system. However, it was not alone: the Stoics proposed a system of propositional logic that was studied by medieval logicians. Also, the problem of multiple generality
was recognized in medieval times. Nonetheless, problems with
syllogistic logic were not seen as being in need of revolutionary
solutions.
Today, some academics claim that Aristotle's system is generally
seen as having little more than historical value (though there is some
current interest in extending term logics), regarded as made obsolete by
the advent of propositional logic and the predicate calculus. Others use Aristotle in argumentation theory to help develop and critically question argumentation schemes that are used in artificial intelligence and legal arguments.
Propositional logic
A propositional calculus or logic (also a sentential calculus) is a
formal system in which formulae representing propositions can be formed
by combining atomic propositions (usually represented with p, q, etc.) using logical connectives ( etc.); these propositions and connectives are the only elements of a standard propositional calculus.
Unlike predicate logic or syllogistic logic where individual subjects
and predicates (which do not have truth values) are the smallest unit,
propositional logic takes full propositions with truth values as its
most basic component. Quantifiers (e.g. or )
are included in extended propositional calculus, but they only quantify
over full propositions, not individual subjects or predicates.
A given propositional logic is a system of formal proof with rules
that establish which well-formed formulae of a given language are
"theorems" by proving them from axioms which are assumed without proof.
Predicate logic
Gottlob Frege's Begriffschrift introduced the notion of quantifier in a graphical notation, which here represents the judgement that is true.
Predicate logic is the generic term for symbolic formal systems such as first-order logic, second-order logic, many-sorted logic, and infinitary logic. It provides an account of quantifiers general enough to express a wide set of arguments occurring in natural language. For example, Bertrand Russell's famous barber paradox, "there is a man who shaves all and only men who do not shave themselves" can be formalised by the sentence , using the non-logical predicate to indicate that x is a man, and the non-logical relation to indicate that x shaves y; all other symbols of the formulae are logical, expressing the universal and existential quantifiers, conjunction, implication, negation and biconditional.
Whilst Aristotelian syllogistic logic specifies a small number of
forms that the relevant part of the involved judgements may take,
predicate logic allows sentences to be analysed into subject and
argument in several additional ways—allowing predicate logic to solve
the problem of multiple generality that had perplexed medieval logicians.
In languages, modality
deals with the phenomenon that sub-parts of a sentence may have their
semantics modified by special verbs or modal particles. For example, "We go to the games" can be modified to give "We should go to the games", and "We can go to the games" and perhaps "We will go to the games".
More abstractly, we might say that modality affects the circumstances
in which we take an assertion to be satisfied. Confusing modality is
known as the modal fallacy.
Aristotle's
logic is in large parts concerned with the theory of non-modalized
logic. Although, there are passages in his work, such as the famous sea-battle argument in De Interpretatione § 9, that are now seen as anticipations of modal logic and its connection with potentiality and time, the earliest formal system of modal logic was developed by Avicenna, who ultimately developed a theory of "temporallymodalized" syllogistic.
While the study of necessity and possibility remained important
to philosophers, little logical innovation happened until the landmark
investigations of C. I. Lewis in 1918, who formulated a family of rival axiomatizations of the alethic modalities. His work unleashed a torrent of new work on the topic, expanding the kinds of modality treated to include deontic logic and epistemic logic. The seminal work of Arthur Prior applied the same formal language to treat temporal logic and paved the way for the marriage of the two subjects. Saul Kripke discovered (contemporaneously with rivals) his theory of frame semantics, which revolutionized the formal technology available to modal logicians and gave a new graph-theoretic way of looking at modality that has driven many applications in computational linguistics and computer science, such as dynamic logic.
Informal reasoning and dialectic
The motivation for the study of logic in ancient times was clear: it
is so that one may learn to distinguish good arguments from bad
arguments, and so become more effective in argument and oratory, and
perhaps also to become a better person. Half of the works of Aristotle's
Organon
treat inference as it occurs in an informal setting, side by side with
the development of the syllogistic, and in the Aristotelian school,
these informal works on logic were seen as complementary to Aristotle's
treatment of rhetoric.
This ancient motivation is still alive, although it no longer takes centre stage in the picture of logic; typically dialectical logic forms the heart of a course in critical thinking,
a compulsory course at many universities. Dialectic has been linked to
logic since ancient times, but it has not been until recent decades
that European and American logicians have attempted to provide
mathematical foundations for logic and dialectic by formalising
dialectical logic. Dialectical logic is also the name given to the special treatment of dialectic in Hegelian and Marxist thought. There have been pre-formal treatises on argument and dialectic, from authors such as Stephen Toulmin (The Uses of Argument), Nicholas Rescher (Dialectics), and van Eemeren and Grootendorst (Pragma-dialectics). Theories of defeasible reasoning
can provide a foundation for the formalisation of dialectical logic and
dialectic itself can be formalised as moves in a game, where an
advocate for the truth of a proposition and an opponent argue. Such
games can provide a formal game semantics for many logics.
Argumentation theory
is the study and research of informal logic, fallacies, and critical
questions as they relate to every day and practical situations. Specific
types of dialogue can be analyzed and questioned to reveal premises,
conclusions, and fallacies. Argumentation theory is now applied in artificial intelligence and law.
Mathematical logic
Mathematical logic comprises two distinct areas of research: the
first is the application of the techniques of formal logic to
mathematics and mathematical reasoning, and the second, in the other
direction, the application of mathematical techniques to the
representation and analysis of formal logic.
The earliest use of mathematics and geometry in relation to logic and philosophy goes back to the ancient Greeks such as Euclid, Plato, and Aristotle. Many other ancient and medieval philosophers applied mathematical ideas and methods to their philosophical claims.
One of the boldest attempts to apply logic to mathematics was the logicism pioneered by philosopher-logicians such as Gottlob Frege and Bertrand Russell. Mathematical theories were supposed to be logical tautologies, and the programme was to show this by means of a reduction of mathematics to logic. The various attempts to carry this out met with failure, from the crippling of Frege's project in his Grundgesetze by Russell's paradox, to the defeat of Hilbert's program by Gödel's incompleteness theorems.
Both the statement of Hilbert's program and its refutation by
Gödel depended upon their work establishing the second area of
mathematical logic, the application of mathematics to logic in the form
of proof theory. Despite the negative nature of the incompleteness theorems, Gödel's completeness theorem, a result in model theory
and another application of mathematics to logic, can be understood as
showing how close logicism came to being true: every rigorously defined
mathematical theory can be exactly captured by a first-order logical
theory; Frege's proof calculus is enough to describe the whole of mathematics, though not equivalent to it.
If proof theory and model theory have been the foundation of
mathematical logic, they have been but two of the four pillars of the
subject. Set theory originated in the study of the infinite by Georg Cantor, and it has been the source of many of the most challenging and important issues in mathematical logic, from Cantor's theorem, through the status of the Axiom of Choice and the question of the independence of the continuum hypothesis, to the modern debate on large cardinal axioms.
Philosophical logic deals with formal descriptions of ordinary, non-specialist ("natural") language,
that is strictly only about the arguments within philosophy's other
branches. Most philosophers assume that the bulk of everyday reasoning
can be captured in logic if a method or methods to translate ordinary
language into that logic can be found. Philosophical logic is
essentially a continuation of the traditional discipline called "logic"
before the invention of mathematical logic. Philosophical logic has a
much greater concern with the connection between natural language and
logic. As a result, philosophical logicians have contributed a great
deal to the development of non-standard logics (e.g. free logics, tense logics) as well as various extensions of classical logic (e.g. modal logics) and non-standard semantics for such logics (e.g. Kripke's supervaluationism in the semantics of logic).
Logic and the philosophy of language are closely related.
Philosophy of language has to do with the study of how our language
engages and interacts with our thinking. Logic has an immediate impact
on other areas of study. Studying logic and the relationship between
logic and ordinary speech can help a person better structure his own
arguments and critique the arguments of others. Many popular arguments
are filled with errors because so many people are untrained in logic and
unaware of how to formulate an argument correctly.
Computational logic
A simple toggling circuit is expressed using a logic gate and a synchronous register.
Logic cut to the heart of computer science as it emerged as a discipline: Alan Turing's work on the Entscheidungsproblem followed from Kurt Gödel's work on the incompleteness theorems.
The notion of the general purpose computer that came from this work was
of fundamental importance to the designers of the computer machinery in
the 1940s.
In the 1950s and 1960s, researchers predicted that when human knowledge could be expressed using logic with mathematical notation,
it would be possible to create a machine that mimics the
problem-solving skills of a human being. This was more difficult than
expected because of the complexity of human reasoning. In the summer of
1956, John McCarthy, Marvin Minsky, Claude Shannon and Nathan Rochester organized a conference on the subject of what they called "artificial intelligence" (a term coined by McCarthy for the occasion). Newell and Simon proudly presented the group with the Logic Theorist and were somewhat surprised when the program received a lukewarm reception.
In logic programming, a program consists of a set of axioms and rules. Logic programming systems such as Prolog compute the consequences of the axioms and rules in order to answer a query.
Today, logic is extensively applied in the field of artificial
intelligence, and this field provide a rich source of problems in formal
and informal logic. Argumentation theory is one good example of how logic is being applied to artificial intelligence. The ACM Computing Classification System in particular regards:
Furthermore, computers can be used as tools for logicians. For
example, in symbolic logic and mathematical logic, proofs by humans can
be computer-assisted. Using automated theorem proving, the machines can find and check proofs, as well as work with proofs too lengthy to write out by hand.
Non-classical logic
The logics discussed above are all "bivalent" or "two-valued"; that is, they are most naturally understood as dividing propositions into true and false propositions. Non-classical logics are those systems that reject various rules of Classical logic.
Hegel developed his own dialectic logic that extended Kant's
transcendental logic but also brought it back to ground by assuring us
that "neither in heaven nor in earth, neither in the world of mind nor
of nature, is there anywhere such an abstract 'either–or' as the
understanding maintains. Whatever exists is concrete, with difference
and opposition in itself".
In 1910, Nicolai A. Vasiliev
extended the law of excluded middle and the law of contradiction and
proposed the law of excluded fourth and logic tolerant to contradiction. In the early 20th century Jan Łukasiewicz
investigated the extension of the traditional true/false values to
include a third value, "possible" (or an indeterminate, a hypothesis) so
inventing ternary logic, the first multi-valued logic in the Western tradition.
A minor modification of the ternary logic was later introduced in a sibling ternary logic model proposed by Stephen Cole Kleene.
Kleene's system differs from the Łukasiewicz's logic with respect to an
outcome of the implication. The former assumes that the operator of implication between two hypotheses produces a hypothesis.
Logics such as fuzzy logic have since been devised with an infinite number of "degrees of truth", represented by a real number between 0 and 1.
Modal logic
is not truth conditional, and so it has often been proposed as a
non-classical logic. However, modal logic is normally formalized with
the principle of the excluded middle, and its relational semantics is bivalent, so this inclusion is disputable.
Controversies
"Is Logic Empirical?"
What is the epistemological status of the laws of logic? What sort of argument is appropriate for criticizing purported principles of logic? In an influential paper entitled "Is Logic Empirical?" Hilary Putnam, building on a suggestion of W. V. Quine,
argued that in general the facts of propositional logic have a similar
epistemological status as facts about the physical universe, for example
as the laws of mechanics or of general relativity,
and in particular that what physicists have learned about quantum
mechanics provides a compelling case for abandoning certain familiar
principles of classical logic: if we want to be realists about the physical phenomena described by quantum theory, then we should abandon the principle of distributivity, substituting for classical logic the quantum logic proposed by Garrett Birkhoff and John von Neumann.
Another paper of the same name by Michael Dummett argues that Putnam's desire for realism mandates the law of distributivity.
Distributivity of logic is essential for the realist's understanding of
how propositions are true of the world in just the same way as he has
argued the principle of bivalence is. In this way, the question, "Is
Logic Empirical?" can be seen to lead naturally into the fundamental
controversy in metaphysics on realism versus anti-realism.
Implication: strict or material
The notion of implication formalized in classical logic does not
comfortably translate into natural language by means of "if ...
then ...", due to a number of problems called the paradoxes of material implication.
The first class of paradoxes involves counterfactuals, such as If the moon is made of green cheese, then 2+2=5, which are puzzling because natural language does not support the principle of explosion. Eliminating this class of paradoxes was the reason for C. I. Lewis's formulation of strict implication, which eventually led to more radically revisionist logics such as relevance logic.
The second class of paradoxes involves redundant premises,
falsely suggesting that we know the succedent because of the antecedent:
thus "if that man gets elected, granny will die" is materially true
since granny is mortal, regardless of the man's election prospects. Such
sentences violate the Gricean maxim of relevance, and can be modelled by logics that reject the principle of monotonicity of entailment, such as relevance logic.
Tolerating the impossible
Georg Wilhelm Friedrich Hegel was deeply critical of any simplified notion of the law of non-contradiction. It was based on Gottfried Wilhelm Leibniz's
idea that this law of logic also requires a sufficient ground to
specify from what point of view (or time) one says that something cannot
contradict itself. A building, for example, both moves and does not
move; the ground for the first is our solar system and for the second
the earth. In Hegelian dialectic, the law of non-contradiction, of
identity, itself relies upon difference and so is not independently
assertable.
Closely related to questions arising from the paradoxes of implication comes the suggestion that logic ought to tolerate inconsistency. Relevance logic and paraconsistent logic are the most important approaches here, though the concerns are different: a key consequence of classical logic and some of its rivals, such as intuitionistic logic, is that they respect the principle of explosion, which means that the logic collapses if it is capable of deriving a contradiction. Graham Priest, the main proponent of dialetheism, has argued for paraconsistency on the grounds that there are in fact, true contradictions.
Rejection of logical truth
The
philosophical vein of various kinds of skepticism contains many kinds
of doubt and rejection of the various bases on which logic rests, such
as the idea of logical form, correct inference, or meaning, typically
leading to the conclusion that there are no logical truths. This is in contrast with the usual views in philosophical skepticism, where logic directs skeptical enquiry to doubt received wisdoms, as in the work of Sextus Empiricus.
Friedrich Nietzsche
provides a strong example of the rejection of the usual basis of logic:
his radical rejection of idealization led him to reject truth as a
"... mobile army of metaphors, metonyms, and anthropomorphisms—in
short ... metaphors which are worn out and without sensuous power; coins
which have lost their pictures and now matter only as metal, no longer
as coins".
His rejection of truth did not lead him to reject the idea of either
inference or logic completely but rather suggested that "logic [came]
into existence in man's head [out] of illogic, whose realm originally
must have been immense. Innumerable beings who made inferences in a way
different from ours perished".
Thus there is the idea that logical inference has a use as a tool for
human survival, but that its existence does not support the existence of
truth, nor does it have a reality beyond the instrumental: "Logic, too,
also rests on assumptions that do not correspond to anything in the
real world".
This position held by Nietzsche however, has come under extreme scrutiny for several reasons. Some philosophers, such as Jürgen Habermas,
claim his position is self-refuting—and accuse Nietzsche of not even
having a coherent perspective, let alone a theory of knowledge. Georg Lukács, in his book The Destruction of Reason,
asserts that, "Were we to study Nietzsche's statements in this area
from a logico-philosophical angle, we would be confronted by a dizzy
chaos of the most lurid assertions, arbitrary and violently
incompatible." Bertrand Russell
described Nietzsche's irrational claims with "He is fond of expressing
himself paradoxically and with a view to shocking conventional readers"
in his book A History of Western Philosophy.