Propositional Logic
Propositional logic, also known as sentential logic and statement logic, is the branch of logic that studies ways of joining and/or modifying entire propositions, statements or sentences to form more complicated propositions, statements or sentences, as well as the logical relationships and properties that are derived from these methods of combining or altering statements. In propositional logic, the simplest statements are considered as indivisible units, and hence, propositional logic does not study those logical properties and relations that depend upon parts of statements that are not themselves statements on their own, such as the subject and predicate of a statement. The most thoroughly researched branch of propositional logic is classical truthfunctional propositional logic, which studies logical operators and connectives that are used to produce complex statements whose truthvalue depends entirely on the truthvalues of the simpler statements making them up, and in which it is assumed that every statement is either true or false and not both. However, there are other forms of propositional logic in which other truthvalues are considered, or in which there is consideration of connectives that are used to produce statements whose truthvalues depend not simply on the truthvalues of the parts, but additional things such as their necessity, possibility or relatedness to one another.
Table of Contents
 Introduction
 History
 The Language of Propositional Logic
 Tautologies, Logical Equivalence and Validity
 Deduction: Rules of Inference and Replacement
 Axiomatic Systems and the Propositional Calculus
 MetaTheoretic Results for the Propositional Calculus
 Other Forms of Propositional Logic
 References and Further Reading
1. Introduction
A statement can be defined as a declarative sentence, or part of a sentence, that is capable of having a truthvalue, such as being true or false. So, for example, the following are statements:
 George W. Bush is the 43rd President of the United States.
 Paris is the capital of France.
 Everyone born on Monday has purple hair.
Sometimes, a statement can contain one or more other statements as parts. Consider for example, the following statement:
 Either Ganymede is a moon of Jupiter or Ganymede is a moon of Saturn.
While the above compound sentence is itself a statement, because it is true, the two parts, “Ganymede is a moon of Jupiter” and “Ganymede is a moon of Saturn”, are themselves statements, because the first is true and the second is false.
The term proposition is sometimes used synonymously with statement. However, it is sometimes used to name something abstract that two different statements with the same meaning are both said to “express”. In this usage, the English sentence, “It is raining”, and the French sentence “Il pleut”, would be considered to express the same proposition; similarly, the two English sentences, “Callisto orbits Jupiter” and “Jupiter is orbited by Callisto” would also be considered to express the same proposition. However, the nature or existence of propositions as abstract meanings is still a matter of philosophical controversy, and for the purposes of this article, the phrases “statement” and “proposition” are used interchangeably.
Propositional logic, also known as sentential logic, is that branch of logic that studies ways of combining or altering statements or propositions to form more complicated statements or propositions. Joining two simpler propositions with the word “and” is one common way of combining statements. When two statements are joined together with “and”, the complex statement formed by them is true if and only if both the component statements are true. Because of this, an argument of the following form is logically valid:
Paris is the capital of France and Paris has a population of over two million.
Therefore, Paris has a population of over two million.
Propositional logic largely involves studying logical connectives such as the words “and” and “or” and the rules determining the truthvalues of the propositions they are used to join, as well as what these rules mean for the validity of arguments, and such logical relationships between statements as being consistent or inconsistent with one another, as well as logical properties of propositions, such as being tautologically true, being contingent, and being selfcontradictory. (These notions are defined below.)
Propositional logic also studies way of modifying statements, such as the addition of the word “not” that is used to change an affirmative statement into a negative statement. Here, the fundamental logical principle involved is that if a given affirmative statement is true, the negation of that statement is false, and if a given affirmative statement is false, the negation of that statement is true.
What is distinctive about propositional logic as opposed to other (typically more complicated) branches of logic is that propositional logic does not deal with logical relationships and properties that involve the parts of a statement smaller than the simple statements making it up. Therefore, propositional logic does not study those logical characteristics of the propositions below in virtue of which they constitute a valid argument:
 George W. Bush is a president of the United States.
 George W. Bush is a son of a president of the United States.
 Therefore, there is someone who is both a president of the United States and a son of a president of the United States.
The recognition that the above argument is valid requires one to recognize that the subject in the first premise is the same as the subject in the second premise. However, in propositional logic, simple statements are considered as indivisible wholes, and those logical relationships and properties that involve parts of statements such as their subjects and predicates are not taken into consideration.
Propositional logic can be thought of as primarily the study of logical operators. A logical operator is any word or phrase used either to modify one statement to make a different statement, or join multiple statements together to form a more complicated statement. In English, words such as “and”, “or”, “not”, “if … then…”, “because”, and “necessarily”, are all operators.
A logical operator is said to be truthfunctional if the truthvalues (the truth or falsity, etc.) of the statements it is used to construct always depend entirely on the truth or falsity of the statements from which they are constructed. The English words “and”, “or” and “not” are (at least arguably) truthfunctional, because a compound statement joined together with the word “and” is true if both the statements so joined are true, and false if either or both are false, a compound statement joined together with the word “or” is true if at least one of the joined statements is true, and false if both joined statements are false, and the negation of a statement is true if and only if the statement negated is false.
Some logical operators are not truthfunctional. One example of an operator in English that is not truthfunctional is the word “necessarily”. Whether a statement formed using this operator is true or false does not depend entirely on the truth or falsity of the statement to which the operator is applied. For example, both of the following statements are true:
 2 + 2 = 4.
 Someone is reading an article in a philosophy encyclopedia.
However, let us now consider the corresponding statements modified with the operator “necessarily”:
 Necessarily, 2 + 2 = 4.
 Necessarily, someone is reading an article in a philosophy encyclopedia.
Here, the first example is true but the second example is false. Hence, the truth or falsity of a statement using the operator “necessarily” does not depend entirely on the truth or falsity of the statement modified.
Truthfunctional propositional logic is that branch of propositional logic that limits itself to the study of truthfunctional operators. Classical (or “bivalent”) truthfunctional propositional logic is that branch of truthfunctional propositional logic that assumes that there are are only two possible truthvalues a statement (whether simple or complex) can have: (1) truth, and (2) falsity, and that every statement is either true or false but not both.
Classical truthfunctional propositional logic is by far the most widely studied branch of propositional logic, and for this reason, most of the remainder of this article focuses exclusively on this area of logic. In addition to classical truthfunctional propositional logic, there are other branches of propositional logic that study logical operators, such as “necessarily”, that are not truthfunctional. There are also “nonclassical” propositional logics in which such possibilities as (i) a proposition’s having a truthvalue other than truth or falsity, (ii) a proposition’s having an indeterminate truthvalue or lacking a truthvalue altogether, and sometimes even (iii) a proposition’s being both true and false, are considered. (For more information on these alternative forms of propositional logic, consult Section VIII below.)
2. History
The serious study of logic as an independent discipline began with the work of Aristotle (384322 BCE). Generally, however, Aristotle’s sophisticated writings on logic dealt with the logic of categories and quantifiers such as “all”, and “some”, which are not treated in propositional logic. However, in his metaphysical writings, Aristotle espoused two principles of great importance in propositional logic, which have since come to be called the Law of Excluded Middle and the Law of Contradiction. Interpreted in propositional logic, the first is the principle that every statement is either true or false, the second is the principle that no statement is both true and false. These are, of course, cornerstones of classical propositional logic. There is some evidence that Aristotle, or at least his successor at the Lyceum, Theophrastus (d. 287 BCE), did recognize a need for the development of a doctrine of “complex” or “hypothetical” propositions, that is, those involving conjunctions (statements joined by “and”), disjunctions (statements joined by “or”) and conditionals (statements joined by “if… then…”), but their investigations into this branch of logic seem to have been very minor.
More serious attempts to study statement operators such as “and”, “or” and “if… then…” were conducted by the Stoic philosophers in the late 3rd century BCE. Since most of their original works—if indeed, these writings were even produced—are lost, we cannot make many definite claims about exactly who first made investigations into what areas of propositional logic, but we do know from the writings of Sextus Empiricus that Diodorus Cronus and his pupil Philo had engaged in a protracted debate about whether the truth of a conditional statement depends entirely on it not being the case that its antecedent (ifclause) is true while its consequent (thenclause) is false, or whether it requires some sort of stronger connection between the antecedent and consequent—a debate that continues to have relevance for modern discussion of conditionals. The Stoic philosopher Chrysippus (roughly 280205 BCE) perhaps did the most in advancing Stoic propositional logic, by marking out a number of different ways of forming complex premises for arguments, and for each, listing valid inference schemata. Chrysippus suggested that the following inference schemata are to be considered the most basic:
 If the first, then the second; but the first; therefore the second.
 If the first, then the second; but not the second; therefore, not the first.
 Not both the first and the second; but the first; therefore, not the second.
 Either the first or the second [and not both]; but the first; therefore, not the second.
 Either the first or the second; but not the second; therefore the first.
Inference rules such as the above correspond very closely to the basic principles in a contemporary system of natural deduction for propositional logic. For example, the first two rules correspond to the rules of modus ponens and modus tollens, respectively. These basic inference schemata were expanded upon by less basic inference schemata by Chrysippus himself and other Stoics, and are preserved in the work of Diogenes Laertius, Sextus Empiricus and later, in the work of Cicero.
Advances on the work of the Stoics were undertaken in small steps in the centuries that followed. This work was done by, for example, the second century logician Galen (roughly 129210 CE), the sixth century philosopher Boethius (roughly 480525 CE) and later by medieval thinkers such as Peter Abelard (10791142) and William of Ockham (12881347), and others. Much of their work involved producing better formalizations of the principles of Aristotle or Chrysippus, introducing improved terminology and furthering the discussion of the relationships between operators. Abelard, for example, seems to have been the first to clearly differentiate exclusive disjunction from inclusive disjunction (discussed below), and to suggest that inclusive disjunction is the more important notion for the development of a relatively simple logic of disjunctions.
The next major step forward in the development of propositional logic came only much later with the advent of symbolic logic in the work of logicians such as Augustus DeMorgan (18061871) and, especially, George Boole (18151864) in the mid19th century. Boole was primarily interested in developing a mathematicalstyle “algebra” to replace Aristotelian syllogistic logic, primarily by employing the numeral “1” for the universal class, the numeral “0” for the empty class, the multiplication notation “xy” for the intersection of classes x and y, the addition notation “x + y” for the union of classes x and y, etc., so that statements of syllogistic logic could be treated in quasimathematical fashion as equations; for example, “No x is y” could be written as “xy = 0”. However, Boole noticed that if an equation such as “x = 1” is read as “x is true”, and “x = 0” is read as “x is false”, the rules given for his logic of classes can be transformed into a logic for propositions, with “x + y = 1” reinterpreted as saying that either x or y is true, and “xy = 1” reinterpreted as meaning that x and y are both true. Boole’s work sparked rapid interest in logic among mathematicians. Later, “Boolean algebras” were used to form the basis of the truthfunctional propositional logics utilized in computer design and programming.
In the late 19th century, Gottlob Frege (18481925) presented logic as a branch of systematic inquiry more fundamental than mathematics or algebra, and presented the first modern axiomatic calculus for logic in his 1879 work Begriffsschrift. While it covered more than propositional logic, from Frege’s axiomatization it is possible to distill the first complete axiomatization of classical truthfunctional propositional logic. Frege was also the first to systematically argue that all truthfunctional connectives could be defined in terms of negation and the material conditional.
In the early 20th century, Bertrand Russell gave a different complete axiomatization of propositional logic, considered on its own, in his 1906 paper “The Theory of Implication”, and later, along with A. N. Whitehead, produced another axiomatization using disjunction and negation as primitives in the 1910 work Principia Mathematica. Proof of the possibility of defining all truth functional operators in virtue of a single binary operator was first published by American logician H. M. Sheffer in 1913, though American logician C. S. Peirce (18391914) seems to have discovered this decades earlier. In 1917, French logician Jean Nicod discovered that it was possible to axiomatize propositional logic using the Sheffer stroke and only a single axiom schema and single inference rule.
The notion of a “truth table” is often utilized in the discussion of truthfunctional connectives (discussed below). It seems to have been at least implicit in the work of Peirce, W. S. Jevons (18351882), Lewis Carroll (18321898), John Venn (18341923), and Allan Marquand (18531924). Truth tables appear explicitly in writings by Eugen Müller as early as 1909. Their use gained rapid popularity in the early 1920s, perhaps due to the combined influence of the work of Emil Post, whose 1921 work makes liberal use of them, and Ludwig Wittgenstein’s 1921 Tractatus LogicoPhilosophicus, in which truth tables and truthfunctionality are prominently featured.
Systematic inquiry into axiomatic systems for propositional logic and related metatheory was conducted in the 1920s, 1930s and 1940s by David Hilbert, Paul Bernays, Alfred Tarski, Jan Łukasiewicz, Kurt Gödel, Alonzo Church, and others. It is during this period, that most of the important metatheoretic results such as those discussed in Section VII were discovered.
Complete natural deduction systems for classical truthfunctional propositional logic were developed and popularized in the work of Gerhard Gentzen in the mid1930s, and subsequently introduced into influential textbooks such as that of F. B. Fitch (1952) and Irving Copi (1953).
Modal propositional logics are the most widely studied form of nontruthfunctional propositional logic. While interest in modal logic dates back to Aristotle, by contemporary standards the first systematic inquiry into this modal propositional logic can be found in the work of C. I. Lewis in 1912 and 1913. Among other wellknown forms of nontruthfunctional propositional logic, deontic logic began with the work of Ernst Mally in 1926, and epistemic logic was first treated systematically by Jaakko Hintikka in the early 1960s. The modern study of threevalued propositional logic began in the work of Jan Łukasiewicz in 1917, and other forms of nonclassical propositional logic soon followed suit. Relevance propositional logic is relatively more recent; dating from the mid1970s in the work of A. R. Anderson and N. D. Belnap. Paraconsistent logic, while having its roots in the work of Łukasiewicz and others, has blossomed into an independent area of research only recently, mainly due to work undertaken by N. C. A. da Costa, Graham Priest and others in the 1970s and 1980s.
3. The Language of Propositional Logic
The basic rules and principles of classical truthfunctional propositional logic are, among contemporary logicians, almost entirely agreed upon, and capable of being stated in a definitive way. This is most easily done if we utilize a simplified logical language that deals only with simple statements considered as indivisible units as well as complex statements joined together by means of truthfunctional connectives. We first consider a language called PL for “Propositional Logic”. Later we shall consider two even simpler languages, PL’ and PL”.
a. Syntax and Formation Rules of PL
In any ordinary language, a statement would never consist of a single word, but would always at the very least consist of a noun or pronoun along with a verb. However, because propositional logic does not consider smaller parts of statements, and treats simple statements as indivisible wholes, the language PL uses uppercase letters ‘‘, ‘‘, ‘‘, etc., in place of complete statements. The logical signs ‘‘, ‘‘, ‘→’, ‘↔’, and ‘‘ are used in place of the truthfunctional operators, “and”, “or”, “if… then…”, “if and only if”, and “not”, respectively. So, consider again the following example argument, mentioned in Section I.
Paris is the capital of France and Paris has a population of over two million.
Therefore, Paris has a population of over two million.
If we use the letter ‘‘ as our translation of the statement “Paris is the captial of France” in PL, and the letter ‘‘ as our translation of the statement “Paris has a population of over two million”, and use a horizontal line to separate the premise(s) of an argument from the conclusion, the above argument could be symbolized in language PL as follows:
In addition to statement letters like ‘‘ and ‘‘ and the operators, the only other signs that sometimes appear in the language PL are parentheses which are used in forming even more complex statements. Consider the English compound sentence, “Paris is the most important city in France if and only if Paris is the capital of France and Paris has a population of over two million.” If we use the letter ‘‘ in language PL to mean that Paris is the most important city in France, this sentence would be translated into PL as follows:
The parentheses are used to group together the statements ‘‘ and ‘‘ and differentiate the above statement from the one that would be written as follows:
This latter statement asserts that Paris is the most important city in France if and only if it is the capital of France, and (separate from this), Paris has a population of over two million. The difference between the two is subtle, but important logically.
It is important to describe the syntax and makeup of statements in the language PL in a precise manner, and give some definitions that will be used later on. Before doing this, it is worthwhile to make a distinction between the language in which we will be discussing PL, namely, English, from PL itself. Whenever one language is used to discuss another, the language in which the discussion takes place is called the metalanguage, and language under discussion is called the object language. In this context, the object language is the language PL, and the metalanguage is English, or to be more precise, English supplemented with certain special devices that are used to talk about language PL. It is possible in English to talk about words and sentences in other languages, and when we do, we place the words or sentences we wish to talk about in quotation marks. Therefore, using ordinary English, I can say that “parler” is a French verb, and “” is a statement of PL. The following expression is part of PL, not English:
However, the following expression is a part of English; in particular, it is the English name of a PL sentence:
“”
This point may seem rather trivial, but it is easy to become confused if one is not careful.
In our metalanguage, we shall also be using certain variables that are used to stand for arbitrary expressions built from the basic symbols of PL. In what follows, the Greek letters ‘‘, ‘‘, and so on, are used for any object language (PL) expression of a certain designated form. For example, later on, we shall say that, if is a statement of PL, then so is . Notice that ‘‘ itself is not a symbol that appears in PL; it is a symbol used in English to speak about symbols of PL. We will also be making use of socalled “Quine corners”, written ‘‘ and ‘‘, which are a special metalinguistic device used to speak about object language expressions constructed in a certain way. Suppose is the statement “” and is the statement ““; then is the complex statement ““.
Let us now proceed to giving certain definitions used in the metalanguage when speaking of the language PL.
Definition: A statement letter of PL is defined as any uppercase letter written with or without a numerical subscript.
Note: According to this definition, ‘‘, ‘‘, ‘‘, ‘‘, and ‘‘ are examples of statement letters. The numerical subscripts are used just in case we need to deal with more than 26 simple statements: in that case, we can use ‘‘ to mean something different than ‘‘, and so forth.
Definition: A connective or operator of PL is any of the signs ‘‘, ‘‘, ‘‘, ‘→’, and ‘↔’.
Definition: A wellformed formula (hereafter abbreviated as wff) of PL is defined recursively as follows:
 Any statement letter is a wellformed formula.
 If is a wellformed formula, then so is .
 If and are wellformed formulas, then so is .
 If and are wellformed formulas, then so is .
 If and are wellformed formulas, then so is .
 If and are wellformed formulas, then so is .
 Nothing that cannot be constructed by successive steps of (1)(6) is a wellformed formula.
Note: According to part (1) of this definition, the statement letters ‘‘, ‘‘ and ‘‘ are wffs. Because ‘‘ and ‘‘ are wffs, by part (3), “” is a wff. Because it is a wff, and ‘‘ is also a wff, by part (6), “” is a wff. It is conventional to regard the outermost parentheses on a wff as optional, so that “” is treated as an abbreviated form of ““. However, whenever a shorter wff is used in constructing a more complicated wff, the parentheses on the shorter wff are necessary.
The notion of a wellformed formula should be understood as corresponding to the notion of a grammatically correct or properly constructed statement of language PL. This definition tells us, for example, that “” is grammatical for PL because it is a wellformed formula, whereas the string of symbols, ““, while consisting entirely of symbols used in PL, is not grammatical because it is not wellformed.
b. Truth Functions and Truth Tables
So far we have in effect described the grammar of language PL. When setting up a language fully, however, it is necessary not only to establish rules of grammar, but also describe the meanings of the symbols used in the language. We have already suggested that uppercase letters are used as complete simple statements. Because truthfunctional propositional logic does not analyze the parts of simple statements, and only considers those ways of combining them to form more complicated statements that make the truth or falsity of the whole dependent entirely on the truth or falsity of the parts, in effect, it does not matter what meaning we assign to the individual statement letters like ‘‘, ‘‘ and ‘‘, etc., provided that each is taken as either true or false (and not both).
However, more must be said about the meaning or semantics, of the logical operators ‘‘, ‘‘, ‘→’, ‘↔’, and ‘‘. As mentioned above, these are used in place of the English words, ‘and’, ‘or’, ‘if… then…’, ‘if and only if’, and ‘not’, respectively. However, the correspondence is really only rough, because the operators of PL are considered to be entirely truthfunctional, whereas their English counterparts are not always used truthfunctionally. Consider, for example, the following statements:
 If Bob Dole is president of the United States in 2004, then the president of the United States in 2004 is a member of the Republican party.
 If Al Gore is president of the United States in 2004, then the president of the United States in 2004 is a member of the Republican party.
For those familiar with American politics, it is tempting to regard the English sentence (1) as true, but to regard (2) as false, since Dole is a Republican but Gore is not. But notice that in both cases, the simple statement in the “if” part of the “if… then…” statement is false, and the simple statement in the “then” part of the statement is true. This shows that the English operator “if… then…” is not fully truthfunctional. However, all the operators of language PL are entirely truthfunctional, so the sign ‘→’, though similar in many ways to the English “if… then…” is not in all ways the same. More is said about this operator below.
Since our study is limited to the ways in which the truthvalues of complex statements depend on the truthvalues of the parts, for each operator, the only aspect of its meaning relevant in this context is its associated truthfunction. The truthfunction for an operator can be represented as a table, each line of which expresses a possible combination of truthvalues for the simpler statements to which the operator applies, along with the resulting truthvalue for the complex statement formed using the operator.
The signs ‘‘, ‘‘, ‘→’, ‘↔’, and ‘‘, correspond, respectively, to the truthfunctions of conjunction, disjunction, material implication, material equivalence, and negation. We shall consider these individually.
Conjunction: The conjunction of two statements and , written in PL as , is true if both and are true, and is false if either is false or is false or both are false. In effect, the meaning of the operator ‘‘ can be displayed according to the following chart, which shows the truthvalue of the conjunction depending on the four possibilities of the truthvalues of the parts:
T
T
F
F T
F
T
F T
F
F
F
Conjunction using the operator ‘‘ is language PL’s rough equivalent of joining statements together with ‘and’ in English. In a statement of the form , the two statements joined together, and , are called the conjuncts, and the whole statement is called a conjunction.
Instead of the sign ‘‘, some other logical works use the signs ‘‘ or ‘‘ for conjunction.
Disjunction: The disjunction of two statements and , written in PL as , is true if either is true or is true, or both and are true, and is false only if both and are false. A chart similar to that given above for conjunction, modified for to show the meaning of the disjunction sign ‘‘ instead, would be drawn as follows:
T
T
F
F T
F
T
F T
T
T
F
This is language PL’s rough equivalent of joining statements together with the word ‘or’ in English. However, it should be noted that the sign ‘‘ is used for disjunction in the inclusive sense. Sometimes when the word ‘or’ is used to join together two English statements, we only regard the whole as true if one side or the other is true, but not both, as when the statement “Either we can buy the toy robot, or we can buy the toy truck; you must choose!” is spoken by a parent to a child who wants both toys. This is called the exclusive sense of ‘or’. However, in PL, the sign ‘‘ is used inclusively, and is more analogous to the English word ‘or’ as it appears in a statement such as (for example, said about someone who has just received a perfect score on the SAT), “either she studied hard, or she is extremely bright”, which does not mean to rule out the possibility that she both studied hard and is bright. In a statement of the form , the two statements joined together, and , are called the disjuncts, and the whole statement is called a disjunction.
Material Implication: This truthfunction is represented in language PL with the sign ‘→’. A statement of the form , is false if is true and is false, and is true if either is false or is true (or both). This truthfunction generates the following chart:
T
T
F
F T
F
T
F T
F
T
T
Because the truth of a statement of the form rules out the possibility of being true and being false, there is some similarity between the operator ‘→’ and the English phrase, “if… then…”, which is also used to rule out the possibility of one statement being true and another false; however, ‘→’ is used entirely truthfunctionally, and so, for reasons discussed earlier, it is not entirely analogous with “if… then…” in English. If is false, then is regarded as true, whether or not there is any connection between the falsity of and the truthvalue of . In a statement of the form , we call the antecedent, and we call the consequent, and the whole statement is sometimes also called a (material) conditional.
The sign ‘‘ is sometimes used instead of ‘→’ for material implication.
Material Equivalence: This truthfunction is represented in language PL with the sign ‘↔’. A statement of the form is regarded as true if and are either both true or both false, and is regarded as false if they have different truthvalues. Hence, we have the following chart:
T
T
F
F T
F
T
F T
F
F
T
Since the truth of a statement of the form requires and to have the same truthvalue, this operator is often likened to the English phrase “…if and only if…”. Again, however, they are not in all ways alike, because ‘↔’ is used entirely truthfunctionally. Regardless of what and are, and what relation (if any) they have to one another, if both are false, is considered to be true. However, we would not normally regard the statement “Al Gore is the President of the United States in 2004 if and only if Bob Dole is the President of the United States in 2004” as true simply because both simpler statements happen to be false. A statement of the form is also sometimes referred to as a (material) biconditional.
The sign ‘‘ is sometimes used instead of ‘↔’ for material equivalence.
Negation: The negation of statement , simply written in language PL, is regarded as true if is false, and false if is true. Unlike the other operators we have considered, negation is applied to a single statement. The corresponding chart can therefore be drawn more simply as follows:
T
FF
T
The negation sign ‘‘ bears obvious similarities to the word ‘not’ used in English, as well as similar phrases used to change a statement from affirmative to negative or viceversa. In logical languages, the signs ‘‘ or ‘‘ are sometimes used in place of ‘‘.
The five charts together provide the rules needed to determine the truthvalue of a given wff in language PL when given the truthvalues of the independent statement letters making it up. These rules are very easy to apply in the case of a very simple wff such as ““. Suppose that ‘‘ is true, and ‘‘ is false; according to the second row of the chart given for the operator, ‘‘, we can see that this statement is false.
However, the charts also provide the rules necessary for determining the truthvalue of more complicated statements. We have just seen that “” is false if ‘‘ is true and ‘‘ is false. Consider a more complicated statement that contains this statement as a part, for example, ““, and suppose once again that ‘‘ is true, and ‘‘ is false, and further suppose that ‘‘ is also false. To determine the truthvalue of this complicated statement, we begin by determining the truthvalue of the internal parts. The statement ““, as we have seen, is false. The other substatement, ““, is true, because ‘‘ is false, and ‘‘ reverses the truthvalue of that to which it is applied. Now we can determine the truthvalue of the whole wff, ““, by consulting the chart given above for ‘→’. Here, the wff “” is our , and “” is our , and since their truthvalues are F and T, respectively, we consult the third row of the chart, and we see that the complex statement “” is true.
We have so far been considering the case in which ‘‘ is true and ‘‘ and ‘‘ are both false. There are, however, a number of other possibilities with regard to the possible truthvalues of the statement letters, ‘‘, ‘‘ and ‘‘. There are eight possibilities altogether, as shown by the following list:
T
T
T
T
F
F
F
F T
T
F
F
T
T
F
F T
F
T
F
T
F
T
F
Strictly speaking, each of the eight possibilities above represents a different truthvalue assignment, which can be defined as a possible assignment of truthvalues T or F to the different statement letters making up a wff or series of wffs. If a wff has n distinct statement letters making up, the number of possible truthvalue assignments is 2^{n}. With the wff, ““, there are three statement letters, ‘‘, ‘‘ and ‘‘, and so there are 8 truthvalue assignments.
It then becomes possible to draw a chart showing how the truthvalue of a given wff would be resolved for each possible truthvalue assignment. We begin with a chart showing all the possible truthvalue assignments for the wff, such as the one given above. Next, we write out the wff itself on the top right of our chart, with spaces between the signs. Then, for each, truthvalue assignment, we repeat the appropriate truthvalue, ‘T’, or ‘F’, underneath the statement letters as they appear in the wff. Then, as the truthvalues of those wffs that are parts of the complete wff are determined, we write their truthvalues underneath the logical sign that is used to form them. The final column filled in shows the truthvalue of the entire statement for each truthvalue assignment. Given the importance of this column, we highlight it in some way. Here, we highlight it in yellow.
 → T
T
T
T
F
F
F
F T
T
F
F
T
T
F
F T
F
T
F
T
F
T
F T
T
T
T
F
F
F
F T
T
F
F
F
F
F
F T
T
F
F
T
T
F
F F
T
T
T
T
T
T
T F
T
F
T
F
T
F
T T
F
T
F
T
F
T
F
Charts such as the one given above are called truth tables. In classical truthfunctional propositional logic, a truth table constructed for a given wff in effects reveals everything logically important about that wff. The above chart tells us that the wff “” can only be false if ‘‘, ‘‘ and ‘‘ are all true, and is true otherwise.
c. Definability of the Operators and the Languages PL’ and PL”
The language PL, as we have seen, contains operators that are roughly analogous to the English operators ‘and’, ‘or’, ‘if… then…’, ‘if and only if’, and ‘not’. Each of these, as we have also seen, can be thought of as representing a certain truthfunction. It might be objected however, that there are other methods of combining statements together in which the truthvalue of the statement depends wholly on the truthvalues of the parts, or in other words, that there are truthfunctions besides conjunction, (inclusive) disjunction, material implication, material equivalence and negation. For example, we noted earlier that the sign ‘‘ is used analogously to ‘or’ in the inclusive sense, which means that language PL has no simple sign for ‘or’ in the exclusive sense. It might be thought, however, that the langauge PL is incomplete without the addition of an additional symbol, say ‘‘, such that would be regarded as true if is true and is false, or is false and is true, but would be regarded as false if either both and are true or both and are false.
However, a possible response to this objection would be to make note that while language PL does not include a simple sign for this exclusive sense of disjunction, it is possible, using the symbols that are included in PL, to construct a statement that is true in exactly the same circumstances. Consider, for example, a statement of the form . It is easily shown, using a truth table, that any wff of this form would have the same truthvalue as a wouldbe statement using the operator ‘‘. See the following chart:
 ↔ T
T
F
F T
F
T
F F
T
T
F T
T
F
F T
F
F
T T
F
T
F
Here we see that a wff of the form is true if either or is true but not both. This shows that PL is not lacking in any way by not containing a sign ‘‘. All the work that one would wish to do with this sign can be done using the signs ‘↔’ and ‘‘. Indeed, one might claim that the sign ‘‘ can be defined in terms of the signs ‘↔’, and ‘‘, and then use the form as an abbreviation of a wff of the form , without actually expanding the primitive vocabulary of language PL.
The signs ‘‘, ‘‘, ‘→’, ‘↔’ and ‘‘, were chosen as the operators to include in PL because they correspond (roughly) the sorts of truthfunctional operators that are most often used in ordinary discourse and reasoning. However, given the preceding discussion, it is natural to ask whether or not some operators on this list can be defined in terms of the others. It turns out that they can. In fact, if for some reason we wished our logical language to have a more limited vocabulary, it is possible to get by using only the signs ‘‘ and ‘→’, and define all other possible truthfunctions in virtue of them. Consider, for example, the following truth table for statements of the form :
 → T
T
F
F T
F
T
F T
F
F
F T
T
F
F F
T
T
T F
T
F
T T
F
T
F
We can see from the above that a wff of the form always has the same truthvalue as the corresponding statement of the form . This shows that the sign ‘‘ can in effect be defined using the signs ‘‘ and ‘→’.
Next, consider the truth table for statements of the form :
 → T
T
F
F T
F
T
F F
F
T
T T
T
F
F T
T
T
F T
F
T
F
Here we can see that a statement of the form always has the same truthvalue as the corresponding statement of the form . Again, this shows that the sign ‘‘ could in effect be defined using the signs ‘→’ and ‘‘.
Lastly, consider the truth table for a statement of the form :
 → → → T
T
F
F T
F
T
F T
F
F
T T
T
F
F T
F
T
T T
F
T
F F
T
T
F F
F
T
F T
F
T
F T
T
F
T T
T
F
F
From the above, we see that a statement of the form always has the same truthvalue as the corresponding statement of the form . In effect, therefore, we have shown that the remaining operators of PL can all be defined in virtue of ‘→’, and ‘‘, and that, if we wished, we could do away with the operators, ‘‘, ‘‘ and ‘↔’, and simply make do with those equivalent expressions built up entirely from ‘→’ and ‘‘.
Let us call the language that results from this simplication PL’. While the definition of a statement letter remains the same for PL’ as for PL, the definition of a wellformed formula (wff) for PL’ can be greatly simplified. In effect, it can be stated as follows:
Definition: A wellformed formula (or wff) of PL’ is defined recursively as follows:
 Any statement letter is a wellformed formula.
 If is a wellformed formula, then so is .
 If and are wellformed formulas, then so is .
 Nothing that cannot be constructed by successive steps of (1)(3) is a wellformed formula.
Strictly speaking, then, the langauge PL’ does not contain any statements using the operators ‘‘, ‘‘, or ‘↔’. One could however, utilize conventions such that, in language PL’, an expression of the form is to be regarded as a mere abbreviation or shorthand for the corresponding statement of the form , and similarly that expressions of the forms and are to be regarded as abbreviations of expressions of the forms or , respectively. In effect, this means that it is possible to translate any wff of language PL into an equivalent wff of language PL’.
In Section VII, it is proven that not only are the operators ‘‘ and ‘→’ sufficient for defining every truthfunctional operator included in language PL, but also that they are sufficient for defining any imaginable truthfunctional operator in classical propositional logic.
Nevertheless, the choice of ‘‘ and ‘→’ for the primitive signs used in language PL’ is to some extent arbitrary. It would also have been possible to define all other operators of PL (including ‘→’) using the signs ‘‘ and ‘‘. On this approach, would be defined as , would be defined as , and would be defined as . Similarly, we could instead have begun with ‘‘ and ‘‘ as our starting operators. On this way of proceeding, would be defined as , would be defined as , and would be defined as .
There are, as we have seen, multiple different ways of reducing all truthfunctional operators down to two primitives. There are also two ways of reducing all truthfunctional operators down to a single primitive operator, but they require using an operator that is not included in language PL as primitive. On one approach, we utilize an operator written ‘’, and explain the truthfunction corresponding to this sign by means of the following chart:
T
T
F
F T
F
T
F F
T
T
T
Here we can see that a statement of the form is false if both and are true, and true otherwise. For this reason one might read ‘’ as akin to the English expression, “Not both … and …”. Indeed, it is possible to represent this truthfunction in language PL using an expression of the form, . However, since it is our intention to show that all other truthfunctional operators, including ‘‘ and ‘‘ can be derived from ‘’, it is better not to regard the meanings of ‘‘ and ‘‘ as playing a part of the meaning of ‘’, and instead attempt (however counterintuitive it may seem) to regard ‘’ as conceptually prior to ‘‘ and ‘‘.
The sign ‘’ is called the Sheffer stroke, and is named after H. M. Sheffer, who first publicized the result that all truthfunctional connectives could be defined in virtue of a single operator in 1913.
We can then see that the connective ‘‘ can be defined in virtue of ‘’, because an expression of the form generates the following truth table, and hence is equivalent to the corresponding expression of the form :
    T
T
F
F T
F
T
F T
T
F
F F
T
T
T T
F
T
F T
F
F
F T
T
F
F F
T
T
T T
F
T
F
Similarly, we can define the operator ‘‘ using ‘’ by noting that an expression of the form always has the same truthvalue as the corresponding statement of the form :
    T
T
F
F T
F
T
F T
T
F
F F
F
T
T T
T
F
F T
T
T
F T
F
T
F F
T
F
T T
F
T
F
The following truth table shows that a statement of the form always has the same truth table as a statement of the form :
   T
T
F
F T
F
T
F T
T
F
F T
F
T
T T
F
T
F F
T
F
T T
F
T
F
Although far from intuitively obvious, the following table shows that an expression of the form always has the same truthvalue as the corresponding wff of the form :
 

 

TT
F
F T
F
T
F T
T
F
F F
F
T
T T
T
F
F T
T
T
F T
F
T
F F
T
F
T T
F
T
F T
F
F
T T
T
F
F F
T
T
T T
F
T
F
This leaves only the sign ‘‘, which is perhaps the easiest to define using ‘’, as clearly , or, roughly, “not both and “, has the opposite truthvalue from itself:
 
TF
T
F F
T T
F
If, therefore, we desire a language for use in studying propositional logic that has as small a vocabulary as possible, we might suggest using a language that employs the sign ‘’ as its sole primitive operator, and defines all other truthfunctional operators in virtue of it. Let us call such a language PL”. PL” differs from PL and PL’ only in that its definition of a wellformed formula can be simplified even further:
Definition: A wellformed formula (or wff) of PL” is defined recursively as follows:
 Any statement letter is a wellformed formula.
 If and are wellformed formulas, then so is .
 Nothing that cannot be constructed by successive steps of (1)(2) is a wellformed formula.
In language PL”, strictly speaking, ‘’ is the only operator. However, for reasons that should be clear from the above, any expression from language PL that involves any of the operators ‘‘, ‘‘, ‘‘, ‘→’, or ‘↔’ could be translated into language PL” without the loss of any of its important logical properties. In effect, statements using these signs could be regarded as abbreviations or shorthand expressions for wffs of PL” that only use the operator ‘’.
Even here, the choice of ‘’ as the sole primitive is to some extent arbitrary. It would also be possible to reduce all truthfunctional operators down to a single primitive by making use of a sign ‘‘, treating it as roughly equivalent to the English expression, “neither … nor …”, so that the corresponding chart would be drawn as follows:
T
T
F
F T
F
T
F F
F
F
T
If we were to use ‘‘ as our sole operator, we could again define all the others. would be defined as ; would be defined as ; would be defined as ; and similarly for the other operators. The sign ‘‘ is sometimes also referred to as the Sheffer stroke, and is also called the Peirce/Sheffer dagger.
Depending on one’s purposes in studying propositional logic, sometimes it makes sense to use a rich language like PL with more primitive operators, and sometimes it makes sense to use a relatively sparse language such as PL’ or PL” with fewer primitive operators. The advantage of the former approach is that it conforms better with our ordinary reasoning and thinking habits; the advantage of the latter is that it simplifies the logical language, which makes certain interesting results regarding the deductive systems making use of the language easier to prove.
For the remainder of this article, we shall primarily be concerned with the logical properties of statements formed in the richer language PL. However, we shall consider a system making use of language PL’ in some detail in Section VI, and shall also make brief mention of a system making use of language PL”.
4. Tautologies, Logical Equivalence and Validity
Truthfunctional propositional logic concerns itself only with those ways of combining statements to form more complicated statements in which the truthvalues of the complicated statements depend entirely on the truthvalues of the parts. Owing to this, all those features of a complex statement that are studied in propositional logic derive from the way in which their truthvalues are derived from those of their parts. These features are therefore always represented in the truth table for a given statement.
Some complex statements have the interesting feature that they would be true regardless of the truthvalues of the simple statements making them up. A simple example would be the wff ““; that is, “ or not “. It is fairly easy to see that this statement is true regardless of whether ‘‘ is true or ‘‘ is false. This is also shown by its truth table:
 TF
T
F T
T F
T T
F
There are, however, statements for which this is true but it is not so obvious. Consider the wff, ““. This wff also comes out as true regardless of the truthvalues of ‘‘, ‘‘ and ‘‘.
 → → → T
T
T
T
F
F
F
F T
T
F
F
T
T
F
F T
F
T
F
T
F
T
F T
F
T
F
T
F
T
F T
T
T
T
T
T
T
T T
T
T
T
F
F
F
F T
T
F
F
T
T
T
T T
T
F
F
T
T
F
F T
T
T
F
T
T
T
T F
F
T
F
F
F
T
F T
F
T
F
T
F
T
F T
T
F
T
T
T
F
T T
T
F
F
T
T
F
F
Statements that have this interesting feature are called tautologies. Let us define this notion precisely.
Definition: a wff is a tautology if and only if it is true for all possible truthvalue assignments to the statement letters making it up.
Tautologies are also sometimes called logical truths or truths of logic because tautologies can be recognized as true solely in virtue of the principles of propositional logic, and without recourse to any additional information.
On the other side of the spectrum from tautologies are statements that come out as false regardless of the truthvalues of the simple statements making them up. A simple example of such a statement would be the wff ““; clearly such a statement cannot be true, as it contradicts itself. This is revealed by its truth table:
 TF
T
F F
F F
T T
F
To state this precisely:
Definition: a wff is a selfcontradiction if and only if it is false for all possible truthvalue assignments to the statement letters making it up.
Another, more interesting, example of a selfcontradiction is the statement ““; this is not as obviously selfcontradictory. However, we can see that it is when we consider its truth table:
 → → TT
F
F T
F
T
F F
T
F
F T
T
F
F T
F
T
T T
F
T
F F
F
F
F F
F
T
F T
F
T
F T
T
F
T T
T
F
F
A statement that is neither selfcontradictory nor tautological is called a contingent statement. A contingent statement is true for some truthvalue assignments to its statement letters and false for others. The truth table for a contingent statement reveals which truthvalue assignments make it come out as true, and which make it come out as false. Consider the truth table for the statement ““:
 → → TT
F
F T
F
T
F T
T
F
F T
F
T
T T
F
T
F F
F
T
T T
T
F
F F
T
T
T F
T
F
T T
F
T
F
We can see that of the four possible truthvalue assignments for this statement, two make it come as true, and two make it come out as false. Specifically, the statement is true when ‘‘ is false and ‘‘ is true, and when ‘‘ is false and ‘‘ is false, and the statement is false when ‘‘ is true and ‘‘ is true and when ‘‘ is true and ‘‘ is false.
Truth tables are also useful in studying logical relationships that hold between two or more statements. For example, two statements are said to be consistent when it is possible for both to be true, and are said to be inconsistent when it is not possible for both to be true. In propositional logic, we can make this more precise as follows.
Definition: two wffs are consistent if and only if there is at least one possible truthvalue assignment to the statement letters making them up that makes both wffs true.
Definition: two wffs are inconsistent if and only if there is no truthvalue assignment to the statement letters making them up that makes them both true.
Whether or not two statements are consistent can be determined by means of a combined truth table for the two statements. For example, the two statements, “” and “” are consistent:
 ↔ TT
F
F T
F
T
F T
T
F
F T
T
T
F T
F
T
F T
F
F
T T
T
F
F F
T
T
F F
T
F
T T
F
T
F
Here, we see that there is one truthvalue assignment, that in which both ‘‘ and ‘‘ are true, that makes both “” and “” true. However, the statements “” and “” are inconsistent, because there is no truthvalue assignment in which both come out as true.
 → TT
F
F T
F
T
F T
T
F
F T
F
T
T T
F
T
F T
F
F
F T
T
F
F F
T
F
F T
F
T
F T
F
T
T F
F
T
T T
T
F
F
Another relationship that can hold between two statements is that of having the same truthvalue regardless of the truthvalues of the simple statements making them up. Consider a combined truth table for the wffs “” and ““:
 → TT
F
F T
F
T
F F
F
T
T T
T
F
F T
T
F
T F
T
F
T T
F
T
F T
T
F
T T
F
T
F F
F
T
F F
F
T
T T
T
F
F
Here we see that these two statements necessarily have the same truthvalue.
Definition: two statements are said to be logically equivalent if and only if all possible truthvalue assignments to the statement letters making them up result in the same resulting truthvalues for the whole statements.
The above statements are logically equivalent. However, the truth table given above for the statements “” and “” show that they, on the other hand, are not logically equivalent, because they differ in truthvalue for three of the four possible truthvalue assignments.
Finally, and perhaps most importantly, truth tables can be utilized to determine whether or not an argument is logically valid. In general, an argument is said to be logically valid whenever it has a form that makes it impossible for the conclusion to be false if the premises are true. (See the encyclopedia article on “Validity and Soundness“.) In classical propositional logic, we can give this a more precise characterization.
Definition: a wff is said to be a logical consequence of a set of wffs , if and only if there is no truthvalue assignment to the statement letters making up these wffs that makes all of true but does not make true.
An argument is logically valid if and only if its conclusion is a logical consequence of its premises. If an argument whose conclusion is and whose only premise is is logically valid, then is said to logically imply .
For example, consider the following argument:
We can test the validity of this argument by constructing a combined truth table for all three statements.
 → → TT
F
F T
F
T
F T
T
F
F T
F
T
T T
F
T
F F
T
F
T T
F
T
F T
T
T
F T
T
F
F T
F
T
F
Here we see that both premises come out as true in the case in which both ‘‘ and ‘‘ are true, and in which ‘‘ is false but ‘‘ is true. However, in those cases, the conclusion is also true. It is possible for the conclusion to be false, but only if one of the premises is false as well. Hence, we can see that the inference represented by this argument is truthpreserving. Contrast this with the following example:
Consider the truthvalue assignment making both ‘‘ and ‘‘ true. If we were to fill in that row of the truthvalue for these statements, we would see that “” comes out as true, but “” comes out as false. Even if ‘‘ and ‘‘ are not actually both true, it is possible for them to both be true, and so this form of reasoning is not truthpreserving. In other words, the argument is not logically valid, and its premise does not logically imply its conclusion.
One of the most striking features of truth tables is that they provide an effective procedure for determining the logical truth, or tautologyhood of any single wff, and for determining the logical validity of any argument written in the language PL. The procedure for constructing such tables is purely rote, and while the size of the tables grows exponentially with the number of statement letters involved in the wff(s) under consideration, the number of rows is always finite and so it is in principle possible to finish the table and determine a definite answer. In sum, classical propositional logic is decidable.
5. Deduction: Rules of Inference and Replacement
a. Natural Deduction
Truth tables, as we have seen, can theoretically be used to solve any question in classical truthfunctional propositional logic. However, this method has its drawbacks. The size of the tables grows exponentially with the number of distinct statement letters making up the statements involved. Moreover, truth tables are alien to our normal reasoning patterns. Another method for establishing the validity of an argument exists that does not have these drawbacks: the method of natural deduction. In natural deduction an attempt is made to reduce the reasoning behind a valid argument to a series of steps each of which is intuitively justified by the premises of the argument or previous steps in the series.
Consider the following argument stated in natural language:
Either cat fur or dog fur was found at the scene of the crime. If dog fur was found at the scene of the crime, officer Thompson had an allergy attack. If cat fur was found at the scene of the crime, then Macavity is responsible for the crime. But officer Thompson didn’t have an allergy attack, and so therefore Macavity must be responsible for the crime.
The validity of this argument can be made more obvious by representing the chain of reasoning leading from the premises to the conclusion:
 Either cat fur was found at the scene of the crime, or dog fur was found at the scene of the crime. (Premise)
 If dog fur was found at the scene of the crime, then officer Thompson had an allergy attack. (Premise)
 If cat fur was found at the scene of the crime, then Macavity is responsible for the crime. (Premise)
 Officer Thompson did not have an allergy attack. (Premise)
 Dog fur was not found at the scene of the crime. (Follows from 2 and 4.)
 Cat fur was found at the scene of the crime. (Follows from 1 and 5.)
 Macavity is responsible for the crime. (Conclusion. Follows from 3 and 6.)
Above, we do not jump directly from the premises to the conclusion, but show how intermediate inferences are used to ultimately justify the conclusion by a stepbystep chain. Each step in the chain represents a simple, obviously valid form of reasoning. In this example, the form of reasoning exemplified in line 5 is called modus tollens, which involves deducing the negation of the antecedent of a conditional from the conditional and the negation of its consequent. The form of reasoning exemplified in step 5 is called disjunctive syllogism, and involves deducing one disjunct of a disjunction on the basis of the disjunction and the negation of the other disjunct. Lastly, the form of reasoning found at line 7 is called modus ponens, which involves deducing the truth of the consequent of a conditional given truth of both the conditional and its antecedent. “Modus ponens” is Latin for affirming mode, and “modus tollens” is Latin for denying mode.
A system of natural deduction consists in the specification of a list of intuitively valid rules of inference for the construction of derivations or stepbystep deductions. Many equivalent systems of deduction have been given for classical truthfunctional propositional logic. In what follows, we sketch one system, which is derived from the popular textbook by Irving Copi (1953). The system makes use of the language PL.
b. Rules of Inference
Here we give a list of intuitively valid rules of inference. The rules are stated in schematic form. Any inference in which any wff of language PL is substituted unformly for the schematic letters in the forms below constitutes an instance of the rule.
Modus ponens (MP):
(Modus ponens is sometimes also called “modus ponendo ponens”, “detachment” or a form of “→elimination”.)
Modus tollens (MT):
(Modus tollens is sometimes also called “modus tollendo tollens” or a form of “→elimination”.)
Disjunctive syllogism (DS): (two forms)
(Disjunctive syllogism is sometimes also called “modus tollendo ponens” or “elimination”.)
Addition (Add): (two forms)
(Addition is sometimes also called “disjunction introduction” or “–introduction”.)
Simplification (Simp): (two forms)
(Simplification is sometimes also called “conjunction elimination” or “elimination”.)
Conjunction (Conj):
(Conjunction is sometimes also called “conjunction introduction”, “introduction” or “logical multiplication”.)
Hypothetical syllogism (HS):
(Hypothetical syllogism is sometimes also called “chain reasoning” or “chain deduction”.)
Constructive dilemma (CD):
Absorption (Abs):
c. Rules of Replacement
The nine rules of inference listed above represent ways of inferring something new from previous steps in a deduction. Many systems of natural deduction, including those initially designed by Gentzen, consist entirely of rules similar to the above. If the language of a system involves signs introduced by definition, it must also allow the substitution of a defined sign for the expression used to define it, or vice versa. Still other systems, while not making use of defined signs, allow one to make certain substitutions of expressions of one form for expressions of another form in certain cases in which the expressions in question are logically equivalent. These are called rules of replacement, and Copi’s natural deduction system invokes such rules. Strictly speaking, rules of replacement differ from inference rules, because, in a sense, when a rule of replacement is used, one is not inferring something new but merely stating what amounts to the same thing using a different combination of symbols. In some systems, rules for replacement can be derived from the inference rules, but in Copi’s system, they are taken as primitive.
Rules of replacement also differ from inference rules in other ways. Inference rules only apply when the main operators match the patterns given and only apply to entire statements. Inference rules are also strictly unidirectional: one must infer what is below the horizontal line from what is above and not viceversa. However, replacement rules can be applied to portions of statements and not only to entire statements; moreover, they can be implemented in either direction.
The rules of replacement used by Copi are the following:
Double negation (DN):
is interreplaceable with
(Double negation is also called “elimination”.)
Commutativity (Com): (two forms)
is interreplaceable with
is interreplaceable with
Associativity (Assoc): (two forms)
is interreplaceable with
is interreplaceable with
Tautology (Taut): (two forms)
is interreplaceable with
is interreplaceable with
DeMorgan’s Laws (DM): (two forms)
is interreplaceable with
is interreplaceable with
Transposition (Trans):
is interreplaceable with
(Transposition is also sometimes called “contraposition”.)
Material Implication (Impl):
is interreplaceable with
Exportation (Exp):
is interreplaceable with
Distribution (Dist): (two forms)
is interreplaceable with
is interreplaceable with
Material Equivalence (Equiv): (two forms)
is interreplaceable with
is interreplaceable with
(Material equivalence is sometimes also called “biconditional introduction/elimination” or “↔introduction/elimination”.)
d. Direct Deductions
A direct deduction of a conclusion from a set of premises consists of an ordered sequence of wffs such that each member of the sequence is either (1) a premise, (2) derived from previous members of the sequence by one of the inference rules, (3) derived from a previous member of the sequence by the replacement of a logically equivalent part according to the rules of replacement, and such that the conclusion is the final step of the sequence.
To be even more precise, a direct deduction is defined as an ordered sequence of wffs, , such that for each step where i is between 1 and n inclusive, either (1) is a premise, (2) matches the form given below the horizontal line for one of the 9 inference rules, and there are wffs in the sequence prior to matching the forms given above the horizontal line, (3) there is a previous step in the sequence where and differs from at most by matching or containing a part that matches one of the forms given for one of the 10 replacement rules in the same place in whcih contains the wff of the corresponding form, and such that the conclusion of the argument is .
Using line numbers and the abbreviations for the rules of the system to annotate, the chain of reasoning given above in English, when transcribed into language PL and organized as a direct deduction, would appear as follows:
1. Premise 2. Premise 3. Premise 4. Premise 5. 2,4 MT 6. 1,5 DS 7. 2,6 MP
There is no unique derivation for a given conclusion from a given set of premises. Here is a distinct derivation for the same conclusion from the same premises:
1. Premise 2. Premise 3. Premise 4. Premise 5. 2,3 Conj 6. 1,5 CD 7. 4,6 DS
Consider next the argument:
This argument has six distinct statement letters, and hence constructing a truth table for it would require 64 rows. The table would have 22 columns, thereby requiring 1,408 distinct T/F calculations. Happily, the derivation of the conclusion of the premises using our inference and replacement rules, while far from simple, is relatively less exhausting:
1. Premise 2. Premise 3. Premise 4. 1 Equiv 5. 4 Simp 6. 2,5 HS 7. 3 Impl 8. 6,7 HS 9. 8 Impl 10. 9 DM 11. 10 Dist 12. 11 Simp 13. 12 Com 14. 13 Dist 15. 14 Simp 16. 15 Taut 17. 16 Add 18. 17 Impl
e. Conditional and Indirect Proofs
Together the nine inference rules and ten rules of replacement are sufficient for creating a deduction for any logically valid argument, provided that the argument has at least one premise. However, to cover the limiting case of arguments with no premises, and simply to facillitate certain deductions that would be recondite otherwise, it is also customary to allow for certain methods of deduction other than direct derivation. Specifically, it is customary to allow the proof techniques known as conditional proof and indirect proof.
A conditional proof is a derivation technique used to establish a conditional wff, that is, a wff whose main operator is the sign ‘→’. This is done by constructing a subderivation within a derivation in which the antecedent of the conditional is assumed as a hypothesis. If, by using the inference rules and rules of replacement (and possibly additional subderivations), it is possible to arrive at the consequent, it is permissible to end the subderivation and conclude the truth of the conditional statement within the main derivation, citing the subderivation as a conditional proof, or ‘CP’ for short. This is much clearer by considering the following example argument:
While a direct derivation establishing the validity of this argument is possible, it is easier to establish the validity of this argument using a conditional derivation.
1. Premise 2. Premise 3. Premise 4. Assumption 5. 1,4 MP 6. 2,4 MP 7. 3 Equiv 8. 7 Simp 9. 6,8 MT 10. 5,9 DS 11. 410 CP
Here in order to establish the conditional statement ““, we constructed a subderivation, which is the portion found at lines 410. First, we assumed the truth of ‘‘, and found that with it, we could derive ‘‘. Given the premises, we therefore had shown that if ‘‘ were also true, so would be ‘‘. Therefore, on the basis of the subderivation we were justified in concluding ““. This is the usual methodology used in logic and mathematics for establishing the truth of a conditional statement.
Another common method is that of indirect proof, also known as proof by reductio ad absurdum. (For a fuller discussion, see the article on reductio ad absurdum in the encyclopedia.) In an indirect proof (‘IP’ for short), our goal is to demonstrate that a certain wff is false on the basis of the premises. Again, we make use of a subderivation; here, we begin by assuming the opposite of that which we’re trying to prove, that is, we assume that the wff is true. If on the basis of this assumption, we can demonstrate an obvious contradiction, that is, a statement of the form , we can conclude that the assumed statement must be false, because anything that leads to a contradiction must be false.
For example, consider the following argument:
While, again, a direct derivation of the conclusion for this argument from the premises is possible, it is somewhat easier to prove that “” is true by showing that, given the premises, it would be impossible for ‘‘ to be true by assuming that it is and showing this to be absurd.
1. Premise 2. Premise 3. Assumption 4. 1,3 MP 5. 2,3 MP 6. 4,5 MP 7. 3,6 Conj 8. 37 IP
Here we were attempting to show that “” was true given the premises. To do this we assumed instead that ‘‘ was true. Since this assumption was impossible, we were justified in concluding that ‘‘ is false, that is, that “” is true.
When making use of either conditional proof or indirect proof, once a subderivation is finished, the lines making it up cannot be used later on in the main derivation or any additional subderivations that may be constructed later on.
This completes our characterization of a system of natural deduction for the language PL.
The system of natural deduction just described is formally adequate in the following sense. Earlier, we defined a valid argument as one in which there is no possible truthvalue assignment to the statement letters making up its premises and conclusion that makes the premises all true but the conclusion untrue. It is provable that an argument in the language of PL is formally valid in that sense if and only if it is possible to construct a derivation of the conclusion of that argument from the premises using the above rules of inference, rules of replacement and techniques of conditional and indirect proof. Space limitations preclude a full proof of this in the metalanguage, although the reasoning is very similar to that given for the axiomatic Propositional Calculus discussed in Sections VI and VII below.
Informally, it is fairly easy to see that no argument for which a deduction is possible in this system could be invalid according to truth tables. Firstly, the rules of inference are all truthpreserving. For example, in the case of modus ponens, it is fairly easy to see from the truth table for any set of statements of the appropriate form that no truthvalue assignment could make both and true while making false. A similar consideration applies for the others. Moreover, truth tables can easily be used to verify that statements of one of the forms mentioned in the rules of replacement are all logically equivalent with those the rule allows one to swap for them. Hence, the statements could never differ in truthvalue for any truthvalue assignment. In case of conditional proof, note that any truthvalue assignment must make either the conditional true, or it must make the antecedent true and consequent false. The antecedent is what is assumed in a conditional proof. So, if the truthvalue assignment makes both it and the premises of the argument true, because the other rules are all truthpreserving, it would be impossible to derive the consequent unless it were also true. A similar consideration justifies the use of indirect proof.
This system represents a useful method for establishing the validity of an argument that has the advantage of coinciding more closely with the way we normally reason. (As noted earlier, however, there are many equivalent systems of natural deduction, all coinciding relatively closely to ordinary reasoning patterns.) One disadvantage this method has, however, is that, unlike truth tables, it does not provide a means for recognizing that an argument is invalid. If an argument is invalid, there is no deduction for it in the system. However, the system itself does not provide a means for recognizing when a deduction is impossible.
Another objection that might be made to the system of deduction sketched above is that it contains more rules and more techniques than it needs to. This leads us directly into our next topic.
6. Axiomatic Systems and the Propositional Calculus
The system of deduction discussed in the previous section is an example of a natural deduction system, that is, a system of deduction for a formal language that attempts to coincide as closely as possible to the forms of reasoning most people actually employ. Natural systems of deduction are typically contrasted with axiomatic systems. Axiomatic systems are minimalist systems; rather than including rules corresponding to natural modes of reasoning, they utilize as few basic principles or rules as possible. Since so few kinds of steps are available in a deduction, relatively speaking, an axiomatic system usually requires more steps for the deduction of a conclusion from a given set of premises as compared to a natural deduction system.
Typically, an axiomatic system consists in the specification of certain wffs that are specified as “axioms”. An axiom is something that is taken as a fundamental truth of the system that does not itself require proof. To allow for the deduction of results from the axioms or the premises of an argument, the system typically also includes at least one (and often only one) rule of inference. Usually, an attempt is made to limit the number of axioms to as few as possible, or at least, limit the number of forms axioms can take.
Because axiomatic systems aim to be minimal, typically they employ languages with simplified vocabularies whenever possible. For classical truthfunctional propositional logic, this might involve using a simpler language such as PL’ or PL” instead of the full language PL.
For most of the remainder of this section, we shall sketch an axiomatic system for classical truthfunctional propositional logic, which we shall dub the Propositional Calculus (or PC for short). The Propositional Calculus makes use of language PL’, described above. That is, the only connectives it uses are ‘→’ and ‘‘, and the other operators, if used at all, would be understood as shorthand abbreviations making use of the definitions discussion in Section III(c).
System PC consists of three axiom schemata, which are forms a wff fits if it is axiom, along with a single inference rule: modus ponens. We make this more precise by specifying certain definitions.
Definition: a wff of language PL’ is an axiom of PC if and only if it is an instance of one of the following three forms:
(Axiom Schema 1, or AS1) (Axiom Schema 2, or AS2) (Axiom Schema 3, or AS3)
Note that according to this definition, every wff of the form is an axiom. This includes an infinite number of different wffs, from simple cases such as ““, to much more complicated cases such as ““.
An ordered stepbystep deduction constitutes a derivation in system PC if and only if each step in the deduction is either (1) a premise of the argument, (2) an axiom, or (3) derived from previous steps by modus ponens. Once again we can make this more precise with the following (more recondite) definition:
Definition: an ordered sequence of wffs is a derivation in system PC of the wff from the premises if and only if, for each wff in the sequence , either (1) is one of the premises , (2) is an axiom of PC, or (3) follows from previous members of the series by the inference rule modus ponens (that is, there are previous members of the sequence, and , such that takes the form ).
For example, consider the following argument written in the language PL’:
The following constitutes a derivation in system PC of the conclusion from the premises:
1. Premise 2. Premise 3. Instance of AS1 4. 1,3 MP 5. 2,4 MP 6. Instance of AS2 7. 5,6 MP 8. 4,7 MP
Historically, the original axiomatic systems for logic were designed to be akin to other axiomatic systems found in mathematics, such as Euclid’s axiomatization of geometry. The goal of developing an axiomatic system for logic was to create a system in which to derive truths of logic making use only of the axioms of the system and the inference rule(s). Those wffs that can be derived from the axioms and inference rule alone, that is, without making use of any additional premises, are called theorems or theses of the system. To make this more precise:
Definition: a wff is said to be a theorem of PC if and only if there is an ordered sequence of wffs, specifically, a derivation, such that, is and each wff in the sequence , is such that either (1) is an axiom of PC, or (2) follows from previous members of the series by modus ponens.
One very simple theorem of system PC is the wff ““. We can show that it is a theorem by constructing a derivation of “” that makes use only of axioms and MP and no additional premises.
1. Instance of AS1 2. Instance of AS1 3. Instance of AS2 4. 2,3 MP 5. 1,4 MP
It is fairly easy to see that not only is “” a theorem of PC, but so is any wff of the form . Whatever happens to be, there will be a derivation in PC of the same form:
1. Instance of AS1 2. Instance of AS1 3. Instance of AS2 4. 2,3 MP 5. 1,4 MP
So even if we make in the above the more complicated wff, for example, ““, a derivation with the same form shows that “” is also a theorem of PC. Hence, we call a theorem schema of PC, because all of its instances are theorems of PC. From now on, let’s call it “Theorem Schema 1”, or “TS1” for short.
The following are also theorem schemata of PC:
(Theorem Schema 2, or TS2) (TS3) (TS4) (TS5)
You may wish to verify this for yourself by attempting to construct the appropriate proofs for each. Be warned that some require quite lengthy derivations!
It is common to use the notation:
to mean that β is a theorem. Similarly, it is common to use the notation:
to mean that it is possible to construct a derivation of making use of as premises.
Considered in terms of number of rules it employs, the axiomatic system PC is far less complex than the system of natural deduction sketched in the previous section. The natural deduction system made use of nine inference rules, ten rules of replacement and two additional proof techniques. The axiomatic system instead, makes use of three axiom schemata and a single inference rule and no additional proof techniques. Yet, the axiomatic system is not lacking in any way.
Indeed, for any argument using language PL’ that is logically valid according to truth tables it is possible to construct a derivation in system PC for that argument. Moreover, every wff of language PL’ that is a logical truth, that is, a tautology according to truth tables, is a theorem of PC. The reverse of these results is true as well; every theorem of PC is a tautology, and every argument for which a derivation in system PC exists is logically valid according to truth tables. These and other features of the Propositional Calculus are discussed, and some are even proven in the next section below.
While the Propositional Calculus is simpler in one way than the natural deduction system sketched in the previous section, in many ways it is actually more complicated to use. For any given argument, a deduction of the conclusion from the premises conducted in PC is likely to be far longer and less psychologically natural than one carried out in a natural deduction system. Such deductions are only simpler in the sense that fewer distinct rules are employed.
System PC is only one of many possible ways of axiomatizing propositional logic. Some systems differ from PC in only very minor ways. For example, we could alter our definition of “axiom” so that a wff is an axiom iff it is an instance of (A1), an instance of (A2), or an instance of the following:
(A3′)
Replacing axiom schema (A3) with (A3′), while altering the way certain deductions must be constructed (making the proofs of many important results longer), has little effect otherwise; the resulting system would have all the same theorems and every argument for which a deduction is possible in the system above would also have a deduction in the revised system, and vice versa.
We also noted above that, strictly speaking, there are an infinite number of axioms of system PC. Instead of utilizing an infinite number of axioms, we might alternatively have utilized only three axioms, namely, the specific wffs:
(A1*)
(A2*)
(A3*)
Note that (A1*) is just a unique wff; on this approach, the wff “” would not count as an axiom, even though it shares a common form with (A1*). To such a system it would be necessary to add an additional inference rule, a rule of substitution or uniform replacement. This would allow one to infer, from a theorem of the system, the result of uniformly replacing any given statement letter (for example, ‘‘ or ‘‘) that occurs within the theorem, with any wff, simple or complex, provided that the same wff replaces all occurrences of the same statement letter in the theorem. On this approach, ““, while not an axiom, would still be a theorem because it could be derived from the rule of uniform replacement twice, that is, by first replacing ‘‘ in (A1*) with ““, and then replacing ‘‘ with ““. The resulting system differs in only subtle ways from our earlier system PC. System PC, strictly speaking, uses only one inference rule, but countenances an infinite number of axioms. This system uses only three axioms, but makes use of an additional rule. System PC, however, avoids this additional inference rule by allowing everything that one could get by substitution in (A1*) to be an axiom. For every theorem , therefore, if is a wff obtained from by uniformly substituting wffs for statement letters in , then is also a theorem of PC, because there would always be a proof of analogous to the proof of only beginning from different axioms.
It is also possible to construct even more austere systems. Indeed, it is possible to utilize only a single axiom schema (or a single axiom plus a rule of replacement). One possibility, suggested by C. A. Meredith (1953), would be to define an axiom as any wff matching the following form:
The resulting system is equally powerful as system PC and has exactly the same set of theorems. However, it is far less psychologically intuitive and straightforward, and deductions even for relatively simple results are often very long.
Historically, the first single axiom schema system made use, instead of language PL’, the even simpler language PL” in which the only connective is the Sheffer stroke, ‘’, as discussed above. In that case, it is possible to make use only of the following axiom schema:
The inference rule of MP is replaced with the rule that from wffs of the form and , one can deduce the wff . This system was discovered by Jean Nicod (1917). Subsequently, a number of possible single axiom systems have been found, some faring better than others in terms of the complexity of the single axiom and in terms of how long deductions for the same results are required to be. (For research in this area, consult McCune et. al. 2002.) Generally, however the more the system allows, the shorter the deductions.
Besides axiomatic and natural deduction forms, deduction systems for propositional logic can also take the form of a sequent calculus; here, rather than specifying definitions of axioms and inference rules, the rules are stated directly in terms of derivability or entailment conditions; for example, one rule might state that if (either or ) then if , then . Sequent calculi, like modern natural deduction systems, were first developed by Gerhard Gentzen. Gentzen’s work also suggests the use of treelike deduction systems rather than linear stepbystep deduction systems, and such tree systems have proven more useful in automated theoremproving, that is, in the creation of algorithms for the mechanical construction of deductions (for example, by a computer). However, rather then exploring the details of these and other rival systems, in the next section, we focus on proving things about the system PC, the axiomatic system treated at length above.
7. Important MetaTheoretic Results for the Propositional Calculus
Note: this section is relatively more technical, and is designed for audiences with some prior background in logic or mathematics. Beginners may wish to skip to the next section.
In this section, we sketch informally the proofs given for certain important features of the Propositional Calculus. Our first topic, however, concerns the language PL’ generally.
Metatheoretic result 1: Language PL’ is expressively adequate, that is, within the context of classical bivalent logic, there are no truthfunctions that cannot be represented in it.
We noted in Section III(c) that the connectives ‘‘, ‘↔’ and ‘‘ can be defined using the connectives of PL’ (‘→’ and ‘‘). More generally, metatheoretic result 1 holds that any statement built using truthfunctional connectives, regardless of what those connectives are, has an equivalent statement formed using only ‘→’ and ‘‘. Here’s the proof.
1. Assume that is some wff built in some language containing any set of truthfunctional connectives, including those not found in PL, PL’ or PL”. For example, might make use of some three or fourplace truthfunctional connectives, or connectives such as the exclusive or, or the sign ‘‘, or any others you might imagine.
2. We need to show that there is a wff formed only with the connectives ‘→’ and ‘‘ that is logically equivalent with . Because we have already shown that forms equivalent to those built from ‘‘, ‘↔’, and ‘‘ can be constructed from ‘→’ and ‘‘, we are entitled to use them as well.
3. In order for it to be logically equivalent to , the wff that we construct must have the same final truthvalue for every possible truthvalue assignment to the statement letters making up , or in other words, it must have the same final column in a truth table.
4. Let be the distinct statement letters making up . For some possible truthvalue assignments to these letters, may be true, and for others may be false. The only hard case would be the one in which is contingent. If were not contingent, it must either be a tautology, or a selfcontradiction. Since clearly tautologies and selfcontradictions can be constructed in PL’, and all tautologies are logically equivalent to one another, and all selfcontradictions are equivalent to one another, in those cases, our job is easy. Let us suppose instead that is contingent.
5. Let us construct a wff in the following way.
(a) Consider in turn each truthvalue assignment to the letters . For each truthvalue assignment, construct a conjunction made up of those letters the truthvalue assignment makes true, along with the negations of those letters the truthvalue assignment makes false. For instance, if the letters involved are ‘‘, ‘‘ and ‘‘, and the truthvalue assignment makes ‘‘ and ‘‘ true but ‘‘ false, consider the conjunction ‘‘.
(b) From the resulting conjunctions, form a complex disjunction formed from those conjunctions formed in step (a) for which the corresponding truthvalue assignment makes true. For example, if the truthvalue assignment making ‘‘ and ‘‘ true but ‘‘ false makes true, include it the disjunction. Suppose, for example, that this truthvalue assignment does make true, as does that assignment in which ‘‘ and ‘‘ and ‘‘ are all made false, but no other truthvalue assignment makes true. In that case, the resulting disjunction would be ‘‘.
6. The wff constructed in step 5 is logically equivalent to . Consider that for those truthvalue assignments making true, one of the conjunctions making up the disjunction is true, and hence the whole disjunction is true as well. For those truthvalue assignments making false, none of the conjunctions making up is true, because each conjunction will contain at least one conjunct that is false on that truthvalue assignment.
7. Because is constructed using only ‘‘, ‘‘ and ‘‘, and these can in turn be defined using only ‘‘ and ‘→’, and because is equivalent to , there is a wff built up only from ‘‘ and ‘→’ that is equivalent to , regardless of the connectives making up .
8. Therefore, PL’ is expressively adequate.
Corollary 1.1: Language PL” is also expressively adequate.
The corollary follows at once from metatheoretic result 1, along with the fact, noted in Section III(c), that ‘→’, and ‘‘ can be defined using only ‘’.
Metatheoretic result 2 (a.k.a. “The Deduction Theorem”): In the Propositional Calculus, PC, whenever it holds that , it also holds that
What this means is that whenever we can prove a given result in PC using a certain number of premises, then it is possible, using all the same premises leaving out one exception, , to prove the conditional statement made up of the removed premise, , as antecedent and the conclusion of the original derivation, , as consequent. The importance of this result is that, in effect, it shows that the technique of conditional proof, typically found in natural deduction (see Section V), is unnecessary in PC, because whenever it is possible to prove the consequent of a conditional by taking the antecedent as an additional premise, a derivation directly for the conditional can be found without taking the antecedent as a premise.
Here’s the proof:
1. Assume that . This means that there is a derivation of in the Propositional Calculus from the premises . This derivation takes the form of an ordered sequence , where the last member of the sequence, , is , and each member of the sequence is either (1) a premise, that is, it is one of , (2) an axiom of PC, (3) derived from previous members of the sequence by modus ponens.
2. We need to show that there is a derivation of , which, while possibly making use of the other premises of the argument, does not make use of . We’ll do this by showing that for each member, , of the sequence of the original derivation: , one can derive without making use of as a premise.
3. Each step in the sequence of the original derivation was gotten at in one of three ways, as mentioned in (1) above. Regardless of which case we are dealing with, we can get the result that . There are three cases to consider:
Case (a): Suppose is a premise of the original argument. Then is either one of or it is itself. In the latter subcase, what we desire to get is that can be gotten at without using as a premise. Because is an instance of TS1, we can get it without using any premises. In the latter case, notice that is one of the premises we’re allowed to use in the new derivation. We’re also allowed to introduce the instance of AS1, . From these, we can get by modus ponens.
Case (b): Suppose is an axiom. We need to show that we can get without using as a premise. In fact, we can get it without using any premises. Because is an axiom, we can use it in the new derivation as well. As in the last case, we have as another axiom (an instance of AS1). From these two axioms, we arrive at by modus ponens.
Case (c): Suppose that was derived from previous members of the sequence by modus ponens. Specifically, there is some and such that both j and k are less than i, and takes the form . We can assume that we have already been able to derive both —that is, —and in the new derivation without making use of . (This may seem questionable in the case that either or was itself gotten at by modus ponens. But notice that this just pushes the assumption back, and eventually one will reach the beginning of the original derivation. The first two steps of the sequence, namely, and , cannot have been derived by modus ponens, since this would require there to have been two previous members of the sequence, which is impossible.) So, in our new derivation, we already have both and .
Notice that is an instance of AS2, and so it can be introduced in the new derivation. By two steps of modus ponens, we arrive at , again without using as a premise.
4. If we continue through each step of the original derivation, showing for each such step , we can get without using as a premise, eventually, we come to the last step of the original derivation, , which is itself. Applying the procedure from step (3), we get that without making use of as a premise. Therefore, the new derivation formed in this way shows that , which is what we were attempting to show.
What’s interesting about this proof for metatheoretic result 2 is that it provides a recipe, given a derivation for a certain result that makes use of one or more premises, for transforming that derivation into one of a conditional statement in which one of the premises of the original argument has become the antecedent. This may be much clearer with an example.
Consider the following derivation for the result that :
1. Premise 2. AS1 3. 1,2 MP 4. AS2 5. 3,4 MP
It is possible to transform the above derivation into one that uses no premises and that shows that is a theorem of PC. The procedure for such a transformation involves looking at each step of the original derivation, and for each one, attempting to derive the same statement, only beginning with ““, without making use of “” as a premise. How this is done depends on whether the step is a premise, an axiom, or a result of modus ponens, and depending on which it is, applying one of the three procedures sketched in the proof above. The result is the following:
1. TS1 2. AS1 3. AS1 4. 2,3 MP 5. AS2 6. 4,5 MP 7. 1,6 MP 8. AS2 9. AS1 10. 8,9 MP 11. AS2 12. 10,11 MP 13. 7,12 MP
The procedure for transforming one sort of derivation into another is purely rote. Moreover, the result is quite often not the most elegant or easy way to show that which you were trying to show. Notice, for example, in the above that lines (2) and (7) are redudant, and more steps were taken than necessary. However, the purely rote procedure is effective.
This metatheoretic result is due to Jacques Herbrand (1930).
It is interesting on its own, especially when one reflects on it as a substitution or replacement for the conditional proof technique. However, it is also very useful for proving other metatheoretic results, as we shall see below.
Metatheoretic result 3: If is a wff of language PL’, and the statement letters making it up are , then if we consider any possible truthvalue assignment to these letters, and consider the set of premises, , that contains if the truthvalue assignment makes true, but contains if the truthvalue assignment makes false, and similarly for , if the truthvalue assignment makes true, then in PC, it holds that , and if it makes false, then .
Here’s the proof.
1. By the definition of a wff, is either itself a statement letter, or ultimately built up from statement letters by the connectives ‘‘ and ‘→’.
2. If is itself a statement letter, then obviously either it or its negation is a member of . It is a member of if the truthvalue assignment makes it true. In that case, obviously, there is a derivation of from , since a premise maybe introduced at any time. If the truthvalue assignment makes it false instead, then is a member of , and so we have a derivation of from , since again a premise may be introduced at any time. This covers the case in which our wff is simply a statement letter.
3. Suppose that is built up from some other wff with the sign ‘‘, that is, suppose that is . We can assume that we have already gotten the desired result for . (Either is a statement letter, in which case the result holds by step (2), or is itself ultimately built up from statement letters, so even if verifying this assumption requires making a similar assumption, ultimately we will get back to statement letters.) That is, if the truthvalue assignment makes true, then we have a derivation of from . If it makes it false, then we have a derivation of from . Suppose that it makes true. Since is the negation of , the truthvalue assignment must make false. Hence, we need to show that there is a derivation of from . Since is , is . If we append to our derivation of from the derivation of , an instance of TS2, we can reach a derivation of by modus ponens, which is what was required. If we assume instead that the truthvalue assignment makes false, then by our assumption, there is a derivation of from . Since is the negation of , this truthvalue assigment must make true. Now, simply is , so we already have a derivation of it from .
4. Suppose instead that is built up from other wffs and with the sign ‘→’, that is, suppose that is . Again, we can assume that we have already gotten the desired result for and . (Again, either they themselves are statement letters or built up in like fashion from statement letters.) Suppose that the truthvalue assignment we are considering makes true. Because is , by the semantics for the sign ‘→’, the truthvalue assignment must make either false or true. Take the first subcase. If it makes false, then by our assumption, there is a derivation of from . If we append to this the derivation of the instance of TS3, , by modus ponens we arrive at derivation of , that is, , from . If instead, the truthvalue assignment makes true, then by our assumption there is a derivation of from . If we add to this derivation the instance of AS1, , by modus ponens, we then again arrive at a derivation of , that is, , from . If instead, the truthvalue assignment makes false, then since is , the truthvalue assignment in question must make true and false. By our assumption, then it is possible to prove both and from . If we concatenate these two derivations, and add to them the derivation of the instance of TS4, , then by two applications of modus ponens, we can derive , which is simply , which is what was desired.
From the above we see that the Propositional Calculus PC can be used to demonstrate the appropriate results for a complex wff if given as premises either the truth or falsity of all its simple parts. This is of course the foundation of truthfunctional logic, that the truth or falsity of those complex statements one can make in it be determined entirely by the truth or falsity of the simple statements entering in to it. Metatheoretic result 3 is again interesting on its own, but it plays a crucial role in the proof of completeness, which we turn to next.
Metatheoretic result 4 (Completeness): If is a wff of language PL’ and a tautology, then is a theorem of the Propositional Calculus.
This feature of the Propositional Calculus is called completeness because it shows that the Propositional Calculus, as a deductive system aiming to capture all the truths of logic, is a success. Every wff true solely in virtue of the truthfunctional nature of the connectives making it up is something that one can prove using only the axioms of PC along with modus ponens. Here’s the proof:
1. Suppose that is a tautology. This means that every possible truthvalue assignment to its statement letters makes it true.
2. Let the statement letters making up be , arranged in some order (say alphabetically and by the numerical order of their subscripts). It follows from (1) and metatheoretic result 3, that there is a derivation in PC of using any possible set of premises that consists, for each statement letter, of either it or its negation.
3. By metatheoretic result 2, we can remove from each of these sets of premises either or , depending on which it contains, and make it an antecedent of a conditional in which is consequent, and the result will be provable without using or as a premise. This means that for every possible set of premises consisting of either or and so on, up until , we can derive both and .
4. The wff is an instance of TS5. Therefore, for any set of premises from which one can derive both and , by two applications of modus ponens, one can also derive itself.
5. Putting (3) and (4) together, we have the result that can be derived from every possible set of premises consisting of either or and so on, up until .
6. We can apply the same reasoning given in steps (3)(5) to remove or its negation from the premise sets by the deduction theorem, arriving at the result that for every set of premises consisting of either or and so on, up until , it is possible to derive . If we continue to apply this reasoning, eventually, we’ll get the result that we can derive with either or its negation as our sole premise. Again, applying the deduction theorem, this means that both and can be proven in PC without using any premises, that is, they are theorems. Concatenating the derivations of these theorems, along with the instance of TS5, , and by two applications of modus ponens, it follows that itself is a theorem, which is what we sought to demonstrate.
The above proof of the completeness of system PC is easier to appreciate when visualized. Suppose, just for the sake of illustration, that the tautology we wish to demonstrate in system PC has three statement letters, ‘‘, ‘‘ and ‘‘. There are eight possible truthvalue assignments to these letters, and since is a tautology, all of them make true. We can sketch in at least this much of ‘s truth table:
 T
T
T
T
F
F
F
F T
T
F
F
T
T
F
F T
F
T
F
T
F
T
F T
T
T
T
T
T
T
T
Now, given this feature of , it follows from metatheoretic result 3, that for every possible combination of premises that consists of either ‘‘ or “” (but not both), either ‘‘ or ““, and ‘‘ or ““, it is possible from those premises to construct a derivation showing . This can be visualized as follows:
By the deduction theorem, we can pull out the last premise from each list of premises and make it an antecedent. However, because from the same remaining list of premises we get both and , we can get by itself from those premises according to TS5. Again, to visualize this:
… and so … and so … and so … and so
We can continue this line of reasoning until all the premises are removed.
and so and so and so and so and so
At the end of this process, we see that is a theorem. Despite only having three axiom schemata and a single inference rule, it is possible to prove any tautology in the simple Propositional Calculus, PC. It is complete in the requisite sense.
This method of proving the completeness of the Propositional Calculus is due to Kalmár (1935).
Corollary 4.1: If a given wff of language PL’ is a logical consequence of a set of wffs , according to their combined truth table, then there is a derivation of with as premises in the Propositional Calculus.
Without going into the details of the proof of this corollary, it follows from the fact that if is a logical consequence of , then the wff of the form is a tautology. As a tautology, it is a theorem of PC, and so if one begins with its derivation in PC and appends a number of steps of modus ponens using as premises, one can derive .
Metatheoretic result 5 (Soundness): If a wff is a theorem of the Propositional Calculus (PC), then is a tautology.
Above, we saw that all tautologies are theorems of PC. The reverse is also true: all theorems of PC are tautologies. Here’s the proof:
1. Suppose that is a theorem of PC. This means that there is an ordered sequence of steps, each of which is either (1) an axiom of PC, or (2) derived from previous members of the sequence by modus ponens, and such that is the last member of the sequence.
2. We can show that not only is a tautology, but so are all the members of the sequence leading to it. The first thing to note is that every axiom of PC is a tautology. To be an axiom of PC, a wff must match one of the axiom schemata AS1, AS2 or AS3. All such wffs must be tautologous; this can easily be verified by constructing truth tables for AS1, AS2 and AS3. (This is left to the reader.)
3. The rule of modus ponens preserves tautologyhood. If is a tautology and is also a tautology, must be a tautology as well. This is because if were not a tautology, it would be false on some truthvalue assignments. However, , as a tautology, is true for all truthvalue assignments. Because a statement of the form is false for any truthvalue assignment making true and false, it would then follow that some truthvalue assignment makes false, which is impossible if it too is a tautology.
4. Hence, we see that the axioms with which we begin the sequence, and every step derived from them using modus ponens, must all be tautologies, and consequently, the last step of the sequence, , must also be a tautology.
This result is called the soundness of the Propositional Calculus; it shows that in it, one cannot demonstrate something that is not logically true.
Corollary 5.1: A wff of language PL’ is a tautology if and only if is a theorem of system PC.
This follows immediately from metatheoretic results 4 and 5.
Corollary 5.2 (Consistency): There is no wff of language PL’ such that both and are theorems of the Propositional Calculus (PC).
Due to metatheoretic result 5, all theorems of PC are tautologies. It is therefore impossible for both and to be theorems, as this would require both to be tautologies. That would mean that both are true for all truthvalue assignments, but obviously, they must have different truthvalues for any given truthvalue assignment, and cannot both be true for any, much less all, such assignments.
This result is called consistency because it guarantees that no theorem of system PC can be inconsistent with any other theorem.
Corollary 5.3: If there is a derivation of the wff with as premises in the Propositional Calculus, then is a logical consequence of the set of wffs , according to their combined truth table.
This is the converse of Corollary 4.1. It follows by the reverse reasoning involved in that corollary. If there is a derivation of taking as premises, then by multiple applications of the deduction theorem (Metatheoretic result 2), it follows that is a theorem of PC. By metatheoretic result 5, must be a tautology. If so, then there cannot be a truthvalue assignment making all of true while making false, and so is a logical consequence of .
Corollary 5.4: There is a derivation of the wff with as premises in the Propositional Calculus if and only if is a logical consequence of , according to their combined truth table.
This follows at once from corollaries 4.1 and 5.3. In sum, then, the Propositional Calculus method of demonstrating something to follow from the axioms of logic is extensionally equivalent to the truth table method of determining whether or not something is a logical truth. Similarly, the truthtable method for testing the validity of an argument is equivalent to the test of being able to construct a derivation for it in the Propositional Calculus. In short, the Propositional Calculus is exactly what we wanted it to be.
Corollary 5.5 (Decidability): The Propositional Calculus (PC) is decidable, that is, there is a finite, effective, rote procedure for determining whether or not a given wff is a theorem of PC or not.
By Corollary 5.1, a wff is a theorem of PC if and only if it is a tautology. Truth tables provide a rote, effective, and finite procedure for determining whether or not a given wff is a tautology. They therefore also provide such a procedure for determining whether or not a given wff is a theorem of PC.
8. Forms of Propositional Logic
So far we have focused only on classical, truthfunctional propositional logic. Its distinguishing features are (1) that all connectives it uses are truthfunctional, that is, the truthvalues of complex statements formed with those connectives depend entirely on the truthvalues of the parts, and (2) that it assumes bivalence: all statements are taken to have exactly one of two truthvalues—truth or falsity—with no statement assigned both truthvalues or neither. Classical truthfunctional propositional logic is the most widely studied and discussed form, but there are other forms of propositional logic.
Perhaps the most well known form of nontruthfunctional propositional logic is modal propositional logic. Modal propositional logic involves introducing operators into the logic involving necessity and possibility, usually along with truthfunctional operators such as ‘→’, ‘‘, ‘‘, etc.. Typically, the sign ‘‘ is used in place of the English operator, “it is necessary that…”, and the sign ‘‘ is used in place of the English operator “it is possible that…”. Sometimes both these operators are taken as primitive, but quite often one is defined in terms of the other, since would appear to be logically equivalent with . (Roughly, it means the same to say that something is not necessarily not true as it does to say that it is possibly true.)
To see that modal propositional logic is not truthfunctional, just consider the following pair of statements:
The first states that it is necessary that . Let us suppose in fact that ‘‘ is true, but might have been false. Since is not necessarily true, the statement “” is false. However, the statement “” is a tautology and so it could not be false. Hence, the statement “” is true. Notice that both ‘‘ and “” are true, but different truthvalues result when the operator ‘‘ is added. So, in modal propositional logic, the truthvalue of a statement does not depend entirely on the truthvalues of the parts.
The study of modal propositional logic involves identifying under what conditions statements involving the operators ‘‘ and ‘‘ should be regarded as true. Different notions or conceptions of necessity lead to different answers to that question. It also involves discovering what inference rules or systems of deduction would be appropriate given the addition of these operators. Here, there is more controversy than with classical truthfunctional logic. For example, in the context of discussions of axiomatic systems for modal propositional logic, very different systems result depending on whether instances of the following schemata are regarded as axiomatic truths, or even truths at all:
If a statement is necessary, is it necessarily necessary? If a statement is possible, is it necessarily possible? A positive answer to the first question is a key assumption in a logical system known as S4 modal logic. Positive answers to both these questions are key assumptions in a logical system known as S5 modal logic. Other systems of modal logic that avoid such assumptions have also been developed. (For an excellent introduction survey, see Hughes and Cresswell 1996.)
Deontic propositional logic and epistemic propositional logic are two other forms of nontruthfunctional propositional logic. The former involves introduction of operators similar to the English operators “it is morally obligatory that…” and “it is morally permissible that…”. Obviously, some things that are in fact true were not morally obligatory, whereas some things that are true were morally obligatory. Again, the truthvalue of a statement in deontic logic does not depend wholly on the truthvalue of the parts. Epistemic logic involves the addition of operators similar to the English operators “it is known that…” and “it is believed that …”. While everything that is known to be the case is in fact the case, not everything that is the case is known to be the case, so a statement built up with a “it is known that…” will not depend entirely on the truth of the proposition it modifies, even if it depends on it to some degree.
Yet another widely studied form of nontruthfunctional propositional logic is relevance propositional logic, which involves the addition of an operator ‘‘ used to connect two statements and to form a statement , which is interpreted to mean that is related to in theme or subject matter. For example, if ‘‘ means that Ben loves Jennifer and ‘‘ means that Jennifer is a pop star, then the statement “” is regarded as true; whereas if ‘‘ means The sun is shining in Tokyo, then “” is false, and hence “” is true. Obviously, whether or not a statement formed using the connective ‘‘ is true does not depend solely on the truthvalue of the propositions involved.
One of the motivations for introducing nontruthfunctional propositional logics is to make up for certain oddities of truthfunctional logic. Consider the truth table for the sign ‘→’ used in Language PL. A statement of the form is regarded as true whenever its antecedent is false or consequent is true. So if we were to translate the English sentence, “if the author of this article lives in France, then the moon is made of cheese” as ““, then strangely, it comes out as true given the semantics of the sign ‘→’ because the antecedent, ‘‘, is false. In modal propositional logic it is possible to define a much stronger sort of operator to use to translate English conditionals as follows:
⥽ is defined as
If we transcribe the English “if the author of this article lives in France, then the moon is made of cheese” instead as “ ⥽ “, then it does not come out as true, because presumably, it is possible for the author of this article to live in France without the moon being made of cheese. Similarly, in relevance logic, one could also define a stronger sort of connective as follows:
is defined as
Here too, if we were to transcribe the English “if the author of this article lives in France, then the moon is made of cheese” as “” instead of simply ““, it comes out as false, because the author of this article living in France is not related to the composition of the moon.
Besides nontruthfunctional logic, other logical systems differ from classical truthfunctional logic by allowing statements to be assigned truthvalues other than truth or falsity, or to be assigned neither truth nor falsity or both truth and falsity. These sorts of logical systems may still be truthfunctional in the sense that the truthvalue of a complex statement may depend entirely on the truthvalues of the parts, but the rules governing such truthfunctionality would be more complicated than for classical logic, because it must consider possibilities that classical logic rejects.
Manyvalued or multivalent logics are those that consider more than two truthvalues. They may admit anything from three to an infinite number of possible truthvalues. The simplest sort of manyvalued logic is one that admits three truthvalues, for example, truth, falsity and indeterminancy. It might seem, for example, that certain statements such as statements about the future, or paradoxical statements such as “this sentence is not true” cannot easily be assigned either truth or falsity, and so, it might be concluded, must have an indeterminate truthvalue. The admission of this third truthvalue requires one to expand the truth tables given in Section III(a). There, we gave a truth table for statements formed using the operator ‘→’; in threevalued logic, we have to decide what the truthvalue of a statement of the form is when either or both of and has an indeterminate truthvalue. Arguably, if any component of a statement is indeterminate in truthvalue, then the whole statement is indeterminate as well. This would lead to the following expanded truth table:
T
T
T
I
I
I
F
F
F T
I
F
T
I
F
T
I
F T
I
F
I
I
I
T
I
T
However, we might wish to retain the feature of classical logic that a statement of the form is always true when its antecedent is false or its consequent is true, and hold that it is indeterminate only when its antecedent is indeterminate and its consequent false or when its antecedent is true and its consequent indeterminate, so that its truth table appears:
T
T
T
I
I
I
F
F
F T
I
F
T
I
F
T
I
F T
I
F
T
T
I
T
T
T
Such details will have an effect on the remainders of the logical systems. For example, if an axiomatic or natural deduction system is created, and a desirable feature is that something be provable from no premises if and only if it is a tautology in the sense of being true (and not just not false) for all possible truthvalue assignments, if we make use of the first truth table for ‘→’, then “” should not be provable, because it is indeterminate when ‘‘ is, whereas if we use the second truth table, then “” should be provable, since it is a tautology according to that truth table, that is, it is true regardless of which of the three truthvalues is assigned to ‘‘.
Here we get just a glimpse at the complications created by admitting more than two truthvalues. If more than three are admitted, and possibly infinitely many, then the issues become even more complicated.
Intuitionistic propositional logic results from rejecting the assumption that every statement is true or false, and countenances statements that are neither. The result is a sort of logic very much akin to a threevalued logic, since “neither true nor false”, while strictly speaking the rejection of a truthvalue, can be thought of as though it were a third truthvalue. In intuitionistic logic, the socalled “law of excluded middle,” that is, the law that all statements of the form are true is rejected. This is because intuitionistic logic takes truth to coincide with direct provability, and it may be that certain statements, such as Goldbach’s conjecture in mathematics, are neither provably the case nor provably not the case.
Paraconsistent propositional logic is even more radical, in countenancing statements that are both true and false. Again, depending on the nature of the system, semantic rules have to be given that determine what the truthvalue or truthvalues a complex statement has when its component parts are both true and false. Such decisions determine what sorts of new or restricted rules of inference would apply to the logical system. For example, paraconsistent logics, if not trivial, must restrict the rules of inference allowable in classical truthfunctional logic, because in systems such as those sketched in Sections V and VI above, from a contradiction, that is, a statement of the form , it is possible to deduce any other statement. Consider, for example, the following deduction in the natural deduction system sketched in Section V.
1. Premise 2. 1 Simp 3. 1 Simp 4. 2 Add 5. 3,4 DS
In order to avoid this result, paraconsistent logics must restrict the notion of a valid inference. In order for an inference to be considered valid, not only must it be truthpreserving, that is, that it be impossible to arrive at something untrue when starting with true premises, it must be falsityavoiding, that is, it must be impossible, starting with true premises, to arrive at something that is false. In paraconsistent logic, where a statement can be both true and false, these two requirements do not coincide. The inference rule of disjunctive syllogism, while truthpreserving, is not falsityavoiding. In cases in which its premises are true, its conclusion can still be false; more specifically, provided that at least one of its premises is both true and false, its conclusion can be false.
Other forms of nonclassical propositional logic, and nontruthfunctional propositional logic, continue to be discovered. Obviously any deviance from classical bivalent propositional logic raises complicated logical and philosophical issues that cannot be fully explored here. For more details both on nonclassical logic, and on nontruthfunctional logic, see the recommended reading section.
9. Suggestions for Further Reading
 Anderson, A. R. and N. D. Belnap [and J. M. Dunn]. 1975 and 1992. Entailment. 2 vols. Princeton, NJ: Princeton University Press.
 Bocheński, I. M. 1961. A History of Formal Logic. Notre Dame, Ind.: University of Notre Dame Press.
 Boole, George. 1847. The Mathematical Analysis of Logic. Cambridge: Macmillan.
 Boole, George. 1854. An Investigation into the Laws of Thought. Cambridge: Macmillan.
 Carroll, Lewis. 1958. Symbolic Logic and the Game of Logic. London: Dover.
 Church, Alonzo. 1956. Introduction to Mathematical Logic. Princeton, NJ: Princeton University Press.
 Copi, Irving. 1953. Introduction to Logic. New York: Macmillan.
 Copi, Irving. 1974. Symbolic Logic. 4th ed. New York: Macmillan.
 da Costa, N. C. A. 1974. “On the Theory of Inconsistent Formal Systems,” Notre Dame Journal of Formal Logic 25: 497510.
 De Morgan, Augustus. 1847. Formal Logic. London: Walton and Maberly.
 Fitch, F. B. 1952. Symbolic Logic: An Introduction. New York: Ronald Press.
 Frege, Gottlob. 1879. Begriffsschrift, ene der arithmetischen nachgebildete Formelsprache des reinen Denkens. Halle: L. Nerbert. Published in English as Conceptual Notation, ed. and trans. by Terrell Bynum. Clarendon: Oxford, 1972.
 Frege, Gottlob. 1923. “Gedankengefüge,” Beträge zur Philosophie des deutchen Idealismus 3: 3651. Published in English as “Compound Thoughts,” in The Frege Reader, edited by Michael Beaney. Oxford: Blackwell, 1997.
 Gentzen, Gerhard. 1934. “Untersuchungen über das logische Schließen” Mathematische Zeitschrift 39: 176210, 40531. Published in English as “Investigations into Logical Deduction,” in Gentzen 1969.
 Gentzen, Gerhard. 1969. Collected Papers. Edited by M. E. Szabo. Amsterdam: NorthHolland Publishing.
 Haack, Susan. 1996. Deviant Logic, Fuzzy Logic. Chicago: University of Chicago Press.
 Herbrand, Jacques. 1930. “Recherches sur la théorie de la démonstration,” Travaux de la Société des Sciences et de la Lettres de Varsovie 33: 133160.
 Hilbert, David and William Ackermann. 1950. Principles of Mathematical Logic. New York: Chelsea.
 Hintikka, Jaakko. 1962. Knowledge and Belief: An Introduction to the Logic of the Two Notions. Ithaca: Cornell University Press.
 Hughes, G. E. and M. J. Cresswell. 1996. A New Introduction to Modal Logic. London: Routledge.
 Jevons, W. S. 1880. Studies in Deductive Logic. London: Macmillan.
 Kalmár, L. 1935. “Über die Axiomatisierbarkeit des Aussagenkalküls,” Acta Scientiarum Mathematicarum 7: 22243.
 Kleene, Stephen C. 1952. Introduction to Metamathematics. Princeton, NJ: Van Nostrand.
 Kneale, William and Martha Kneale. 1962. The Development of Logic. Clarendon: Oxford.
 Lewis, C. I. and C. H. Langford. 1932. Symbolic Logic. New York: Dover.
 Łukasiewicz, Jan. 1920. “O logice trojwartosciowej,” Ruch Filozoficny 5: 170171. Published in English as “On ThreeValued Logic,” in Łukasiewicz 1970.
 Łukasiewicz, Jan. 1970. Selected Works. Amsterdam: NorthHolland.
 Łukasiewicz, Jan and Alfred Tarski. 1930. “Untersuchungen über den Aussagenkalkül,” Comptes Rendus des séances de la Société des Sciences et de la Lettres de Varsovie 32: 3050. Published in English as “Investigations into the Sentential Calculus,” in Tarski 1956.
 Mally, Ernst. 1926. Grundgesetze des Sollens: Elemente der Logik des Willens. Graz: Leuschner und Lubensky.
 McCune, William, Robert Veroff, Branden Fitelson, Kenneth Harris, Andrew Feist and Larry Wos. 2002. “Short Single Axioms for Boolean Algebra,” Journal of Automated Reasoning 29: 116.
 Mendelson, Elliot. 1997. Introduction to Mathematical Logic. 4th ed. London: Chapman and Hall.
 Meredith, C. A. 1953. “Single Axioms for the Systems (C, N), (C, O) and (A, N) of the Twovalued Propositional Calculus,” Journal of Computing Systems 3: 15562.
 Müller, Eugen, ed. 1909. Abriss der Algebra der Logik, by E. Schröder. Leipzig: Teubner.
 Nicod, Jean. 1917. “A Reduction in the Number of the Primitive Propositions of Logic,” Proceedings of the Cambridge Philosophical Society 19: 3241.
 Peirce, C. S. 1885. “On the Algebra of Logic,” American Journal of Mathematics 7: 180202.
 Post, Emil. 1921. “Introduction to a General Theory of Propositions,” American Journal of Mathematics 43: 163185.
 Priest, Graham, Richard Routley and Jean Norman, eds. 1990. Paraconsistent Logic. Munich: Verlag.
 Prior, Arthur. 1990. Formal Logic. 2nd. ed. Oxford: Oxford University Press.
 Read, Stephen, 1988. Relevant Logic. New York: Blackwell.
 Rescher, Nicholas. 1966. The Logic of Commands. London: Routledge and Kegan Paul.
 Rescher, Nicholas. 1969. ManyValued Logic. New York: McGraw Hill.
 Rosser, J. B. 1953. Logic for Mathematicians. New York: McGraw Hill.
 Russell, Bertrand. 1906. “The Theory of Implication,” American Journal of Mathematics 28: 159202.
 Schlesinger, G. N. 1985. The Range of Epistemic Logic. Aberdeen: Aberdeen University Press.
 Sheffer, H. M. 1913. “A Set of Five Postulates for Boolean Algebras with Application to Logical Constants,” Transactions of the American Mathematical Society 14: 48188.
 Smullyan, Raymond. 1961. Theory of Formal Systems. Princeton: Princeton University Press.
 Tarski, Alfred. 1956. Logic, Semantics and MetaMathematics. Oxford: Oxford University Press.
 Urquhart, Alasdair. 1986. “Manyvalued Logic,” In Handbook of Philosophical Logic, vol. 3, edited by D. Gabbay and F. Guenthner. Dordrecht: Reidel.
 Venn, John. 1881. Symbolic Logic. London: Macmillan.
 Whitehead, Alfred North and Bertrand Russell. 19101913. Principia Mathematica. 3 vols. Cambridge: Cambridge University Press.
 Wittgenstein, Ludwig. 1922. Tractatus LogicoPhilosophicus. London: Routledge and Kegan Paul.
Author Information
Kevin C. Klement
Email: klement@philos.umass.edu
University of Massachusetts, Amherst
U. S. A.