Elements of Deductive Logic Lecture Note
Elements of Deductive Logic Lecture Note
AC PASEAU
Faculty of Philosophy
Oxford University
Lecture Notes
Hilary 2019
Lecture 1
In an ideal world, we would devote the first couple of lectures to a review of
mathematical notation and basic proof techniques (e.g. argument by contra-
position, argument by contradiction). There’s no time for that, so I refer you
to Stephen Blamey’s notes. That said, it’s worth mentioning notation we’ll
use fairly often: A B means that A is a (proper or improper) subset of B,
in other words that every element of A is an element of B; A Y B tx : x P A
or x P B u is the set of elements in A or B (or both); A X B tx : x P A
and x P B u is the set of elements in both A and B ; and AzB tx : x P A
and x R B u is the set of elements of A not in B. For simplicity, I’ll be casual
about use and mention, as I was in the last sentence by omitting quotation
marks, and I’ll certainly dispense with Quine quotes (if you don’t know what
that means, don’t worry). We’ll also assume knowlege of The Logic Manual,
covered in the Introduction to Logic course.
The main mathematical tool we’ll need is proof by induction. I expect
most of you have seen this before, so I’ll be relatively brief. One stan-
dard form of the PMI (Principle of Mathematical Induction), in premise–
conclusion form, is:
Φp0q
@n Φpnq Ñ Φpn 1q
@nΦpnq
The quantifiers here range over the natural numbers (0, 1, 2, 3, . . .), and Φ is
any numerical property. Observe that this statement of the PMI is informal,
i.e. formulated in the language of informal mathematics. If/when you take a
more advanced course in logic, you’ll come across a formal statement of the
1
I’m grateful to past and present EDL students, especially Joe Deakin, Emma Baldas-
sari, Paolo Marimon and Raymond Douglas, for comments. Thanks also to Beau Madison
Mount and Joel David Hamkins.
1
PMI in an extension of L that contains function symbols. Here’s another
version of the PMI, sometimes called the strong form:
Φp0q
@n @k ¤ nΦpkq Ñ Φpn 1q
@nΦpnq
It’s an easy exercise to show that these two versions of the PMI are equivalent
(and in particular that they are no different in strength). As you’ll see, we’ll
tend to use the latter more than the former. In lectures, I also mentioned the
Least Number Principle, another equivalent of the PMI, which states that
if at least one natural number has property Φ then there is a least natural
number with property Φ.
Let’s now have two examples of the PMI’s use, the first from arithmetic,
the second from logic.
You may have come across non-inductive proofs of this proposition. Our
proof will be by induction.
Proof. First, let’s check that 0 has the property. Yes, it does, since the
sum of the natural numbers from 0 to 0, namely 0, is indeed 0p02 1q 0. So
we’ve checked the base case.
Next, suppose that the sum of the natural numbers from 0 to N is N pN2 1q .
This is the assumption in the induction step. On that assumption, we need
to prove that the sum of the numbers from 0 to N 1 is pN 1qppN 1q 1q
pN 1qppN 2q . Now
2
2
p q
Σ i Σ i N 1 N N 1
N 1 pN 1qp2 N 2q
¤¤
0 i N 1 ¤¤
0 i N 2
The middle one of the three equations follows from the induction hypothesis
and the last one summarises a short algebraic manipulation. Since we’ve
proved the base case and the induction step, we can conclude by the PMI
that all numbers have the stated property.
The next example of a proof by induction is much more typical of the kind of
reasoning used in this course. Before that, though, we need some definitions.
2
Suppose for example that φ is ppP ^ Qq ^ Rq _ P1 . Then Connpφq
t^, _u, N Connpφq 3, and Atompφq SenLettpφq tP, Q, R, P1u. In
fact, Atompφq SenLettpφq more generally for all L1 -sentences φ. Note
that Connpφq is a set of connective types, whereas N Connpφq counts the
number of connective tokens in φ. Many of the proofs in this course show
that all formulas have some property Φ ‘by induction on the complexity of φ’.
This means that the induction is on N Connpφq: we prove that a formula with
no connectives has the property, and that if a formula with n connectives has
the property Ψ then a formula with n 1 connectives also has the property Ψ;
we then conclude that all formulas have the property. That gives us a handy
way of proving facts about all formulas, by associating them with natural
numbers in this fashion. In other words, what we’re doing is proving that 0 is
such that every formula with that many connectives has the property Ψ, and
that every number n is such that every formula with that many connectives
has the property Ψ, then n 1 is also such that every formula with that
many connectives has the property Ψ. So every number n is such that every
formula with that many connectives has the property Ψ; so all formulas have
the property.
Here’s an example of this approach at work.
Proof. We’ll use induction on a stronger claim, for reasons that will
become apparent. Let Φpnq be the claim:
3
Atompφi q, for i 1, 2. Hence by the induction hypothesis we know that
|φ1|A T and |φ2|A T . The truth-table for Ø then yields, for this same
structure A :
Notice two facts about this proof. We had to use the second rather
than the first version of the PMI because the inductive step involves two
formulas φ1 and φ2 whose complexity is unknown. All we know about the
formulas’ complexity is that N Connpφ1 q N Connpφ2 q n, so our inductive
hypothesis must perforce be that the required property holds for formulas of
any complexity ¤ n. The second point is that we proved the proposition from
a stronger assumption: whereas the proposition states that a formula whose
only connective if any is Ø is true in some L1 -structure, our proof showed
that such a formula is true in the structure that maps all the formula’s atoms
to T . If in the inductive step all we could draw on were the hypotheses that
φ1 is true in some structure A1 and that φ2 is true in some structure A2
then we would have been stuck. For without knowing that A1 and A2 agree
on the atoms of φ, we couldn’t use them to define an L1 -structure in which
φ φ1 Ø φ2 is true. Hence the moral: choose your inductive hypotheses
wisely!
At the end of the first lecture, we covered some of the material to follow
in the next section. But I’ll end the section here since it’s a natural break.
4
Lecture 2
Wherein we make a start on metalogic proper and prove the Interpolation
Theorem for propositional logic. Henceforth we’ll use 1 and 0 as truth-values
rather than T and F , as we can algebraically manipulate the former more
easily.
It’s useful to consider a slight extension of L1 which I’ll (unimaginatively)
call L1 . This logic expands the language of L1 with two atomic sentences
J and K such that |J|A 1 and |K|A 0 for all structures A. Note that
J and K are atoms but not sentence letters. So whereas in L1 , SenLettpφq
is always identical to Atompφq, these notions can come apart in L1 ; e.g.
SenLettpP Ñ Jq tP u whereas AtompP Ñ Jq tP, Ju. We’ll assume
that L1 -consequence respects L1 -consequence, i.e. that a subset Γ of L1 -
sentences entails an L1 -sentence φ in L1 iff Γ entails φ in L1 .
There now follows a sequence of lemmas and definitions, of great use for
the rest of the course. Intuitively, the next lemma tells us that the truth-
value of a sentence in a structure is fixed by the truth-value of its sentence
letters in that structure. Although that should seem obvious enough, it does
require proof. The lemma, note, applies to both L1 and L1 .
5
that of ψ, and the truth-value of each of φ1 ^ φ2 and φ1 _ φ2 and φ1 Ñ φ2
and φ1 Ø φ2 is likewise determined by those of φ1 and φ2 , A and B agree on
φ. In other words, |φ|A |φ|B .
P Q P ^Q
1 1 1
1 0 0
0 1 0
0 0 0
6
φ is a tautology just when, for all A, A ( φ. And φ is a contradiction
just when, for all A, A * φ.
Γ ( φ means: for all A if A ( Γ then A ( φ.
φ and ψ are logically equivalent just when: for all A, A ( φ iff A ( ψ.
We note that the symbol ‘(’ is ambiguous in logic. ‘A ( φ’ means that
the structure A satisfies the sentence φ, whereas ‘Γ ( φ’ means that all
structures that satisfy Γ also satisfy φ. In practice, it would be hard to
confuse the two, as both notation and context should make it clear whether
the symbol to the left of ‘(’ denotes a set of sentences or a structure. In
lectures, I also used φ )( ψ to mean that φ and ψ are logically equivalent,
but I won’t do so here.
Let’s now move on to substitution. We’d like to capture the idea that we
can replace all occurrences of a sentence letter, or more generally sequence
of letters, with some formula(s). This motivates the following definition.
Definition 3 Let Sen(L) be the set of logic L’s sentences. Let π : SenLettpL1 q Ñ
SenpL1 q, i.e. π is a function which maps the sentence letters of L1 to sen-
tences of L1 . Then we extend π to a map defined on all the sentences of
L1 which by abuse of notation we also label π and which we represent by a
superscript, as follows:
απ πpαq if α P SenLettpL1 q
Jπ J
Kπ K
p φqπ pφπ q
pφ ^ ψqπ φπ ^ ψπ
pφ _ ψqπ φπ _ ψπ
pφ Ñ ψqπ φπ Ñ ψπ
pφ Ø ψqπ φπ Ø ψπ
We also extend π to sets of sentences in the obvious way: Γπ tγ π : γ P
Γu.
Given such a map π : SenLettpL1 q Ñ SenpL1 q, we define a correspond-
ing map on structures. Given any structure A, let Aπ be the structure defined
by the following stipulation:
|α|A |πpαq|A for all α P SenLettpL1 q
π
The value |φ|A for a complex sentence φ is then fixed by the values of |α|A
π π
for φ’s sentence letters (by the Relevance Lemma). The following lemma
shows that we can ‘move the superscript down from the sentence to the
structure’.
7
Lemma 4 (Substitution Lemma) |φπ |A |φ|A π , for all φ and A.
|φπ |A |p ψqπ |A
| ψ π |A
1 | ψ π |A
1 | ψ |A π
| ψ |A π
|φ|A π
|φ1 ^ φ2|A π
|φ|A π
Corollary 5 If Γ ( φ then Γπ ( φπ .
Corollary 6 Suppose ( φi Ø ψi (i.e. φi and ψi are logical equivalents) for
1 ¤ i ¤ N . Let χpφ1 {α1 , , φN {αN q be the formula obtained by replacing all
8
occurrences (if any) of the sentence letter αi in χ by φi , for 1 ¤ i ¤ N ; and
let χpψ1 {α1 , , ψN {αN q be the formula obtained by replacing all occurrences
(if any) of the sentence letter αi in χ by ψi , for 1 ¤ i ¤ N . Note that
this is simultaneous, not sequential, substitution in the sense of π. Then
( χpφ1{α1, , φN {αN q Ø χpψ1{α1, , ψN {αN q.
We now come to the highlight of this lecture.
Theorem 7 (Interpolation Theorem for L1 ) Suppose φ ( ψ. Then there
is a sentence λ such that: (i) Senlettpλq Senlettpφq X Senlettpψ q; (ii)
φ ( λ; and (iii) λ ( ψ. λ is known as an interpolant for the sequent φ ( ψ.
Proof We may assume that SenlettpφqzSenlettpψ q is non-empty; for if
it’s empty then Senlettpφq Senlettpψ q, in which case we may take φ as
our interpolant. Let’s now define a set Π of substitutions as follows:
Π tπ : π pαq P tJ, Ku for all α P SenlettpφqzSenlettpψ q and π pαq
α if α is a sentence letter not in SenlettpφqzSenlettpψ qu
Intuitively, a substitution π gets rid of the sentence letters in φ but not ψ by
replacing each of them by either J or K. Π then consists of all the possible
ways of doing so: it’s the set of all substitutions of this kind.
Given Π, we define λ as follows:
λ φπ
P
π Π
9
(iii) Suppose A ( λ. Since λ φπ , A ( φπ for some π
P Π, i.e. |φπ |A 1
P
for this π. By the Substitution Lemma, it follows that |φ|A 1 for
π Π
π
It’s worth thinking about the relative strength of interpolants for a given
sequent. How does λ as defined in the theorem’s proof compare to other
interpolants for the sequent φ ( ψ?
The interpolation theorem for L1 is a relatively straightforward corollary
of the interpolation theorem for L1 . To prove it, we require a couple of
further lemmas.
10
The theorem is ‘best possible’ because we can’t remove the restriction to
φ not being a contradiction or ψ not being a tautology. To see this, consider
for example the sequents P ( Q _ Q and P ^ P ( Q.
11
Lecture 3
Last time we proved the Interpolation Theorem for L1 , and, as a corollary,
the Interpolation Theorem for L1 . The topic of today’s lecture is duality.
It’s a fairly self-contained topic, popular with EDL examiners, so well worth
getting under your belt. To introduce it, it will be useful to know something
about truth-functions.
12
The most natural setting for duality is a propositional logic in which there
is a unique way to represent each connective’s dual. In the propositional
logics L1 or L1 , defining a connective’s dual involves some arbitrary choices.
For example, we can define P Ñ Q as pQ Ñ P q, or as P ^ Q, or in
countless other ways. There is no such thing as the ‘right’ definition, only
a choice to be made on pragmatic grounds. To avoid this, and to introduce
you to a third propositional logic, we work in L1 . We’ll define this logic
informally, as we did L1 .
fc pt1 , , tn q 1 fc p1 t1 , , 1 tn q
We note that c exists and that it’s unique. In languages such as L1 which
lack a primitive symbol for some connectives’ duals, we’d have to define c
13
in some arbitrary fashion, as mentioned, and of course the dual map would
only be defined on the connectives the language happens to contain.
Let’s have some examples, starting with negation. (In these examples,
construe A as a variable ranging over L1 -structures.) From our definitions:
| φ|A 1 pf p1 |φ|Aqq 1 p1 p1 |φ|Aqqq 1 |φ|A | φ|A. This
calculation shows that , or in other words that is self-dual.
For our second example, let’s check that ^ _. Here’s the calculation:
f^ p|φ1 |A , |φ2 |A q 1 f^ p1 |φ1 |A , 1 |φ2 |A q 1 p1 |φ1 |A qp1 |φ2 |A q
|φ1|A |φ2|A |φ1|A|φ2|A f_p|φ1|A, |φ2|Aq. Thus ^ _. We may similarly
check that _ ^. In fact, we can deduce that _ ^ from ^ _ and
the lemma to follow.
14
As for the induction step:
cpφ1 , , φn q cpφ1 , , φnq
cpφ
1 , , φn q
cpφ1, , φnq
cpφ1, , φnq
The first two equations are definitional, the third uses the inductive hypoth-
esis and the fourth the fact that c c.
|φ|A |φ|A 1
Proof By induction on the complexity of φ. For the base case, we note
from definitions that if α is a sentence letter then |α |A |α|A 1 |α|A .
Also |J |A |K|A 0 and |K |A |J|A 1, so each of J and K also has
the noted property.
For the induction step, let φ be cpφ1 , , φn q. Then:
|φ|A | cpφ1, , φnq |A
|cpφ1 , , φnq|A
fc p|φ1 |A, , |φn|Aq
1 fcp1 |φ1 |A, , 1 |φn|Aq
1 fcp|φ1|A , , |φn|A q
1 |cpφ1, , φnq|A
1 |φ|A
The main step, from the fourth to the fifth line, uses the induction hypothesis.
This proves the result.
16
Lecture 4
Last time was all about duality. Today, we’ll look at Expressive Adequacy
and the Compactness Theorem. Recall from the first lecture that we can’t
build a contradiction making use of the connective Ø. Thus the set tØu is
not expressively adequate. So which sets of connectives are? First, let’s get
clear on what the question is.
17
χj αjk
¤¤
1 k N
f p|P1 |A , , |Pn |A q 1
õ
jA P T pf q
õ
χjA is a disjunct in φ and |χjA |A 1
õ
|φpP1, , Pnq|A 1
First biconditional: by the definition of jA and T pf q. Second biconditional:
by the definition of φ in terms of the χj and the properties of conjunctions.
Third biconditional, top to bottom: if χjA is a disjunct in φ and is true in
A then clearly so is φ. Third biconditional, bottom to top: if |φ|A 1 then
some disjunct χj in φ must be true in A; by the properties of conjunctions
|χj |A 1 only if j jA.
The proofs of the next three corollaries and proposition are left as exercises.
Corollary 16 The sets t^, u, t_, u and tÑ, u are all expressively ade-
quate.
18
Definition 10 A literal is a sentence letter or its negation. Any disjunction
of conjunctions of literals is in disjunctive normal form and any conjunction
of disjunctions of literals is in conjunctive normal form.
Definition 12 Γ is finitely satisfiable just when all of its finite subsets are
satisfiable.
Proof. We’ll prove the theorem for L1 , L2 and L . In fact, our argument
proves a much more general result: any logic with a sound and complete
procedure is compact. Here’s the deceptively simple argument:
Here Γfin is some finite subset of Γ. Anything deserving of the name of ‘proof
procedure’ usually satisfies a host of syntactic requirements. Given soundness
and completeness the only such requirement needed for the validity of the
19
inference above is that the step from (2) to (3) be valid, i.e. that proofs draw
only on finitely many premisses. The argument just given therefore applies
to any logic which has a sound and complete proof procedure in this sense.
In particular, it includes L1 , L2 and L , which we know from Introduction
to Logic all satisfy this condition, with $ interpreted as ‘provable in ND’.
Sketch Proof. Suppose for reductio that φ were a sentence with this
property. Then φ would be truein all and only finite structures. For each
n ¥ 1, let D¥n be Dx1 Dxn p xi xj q. Clearly, D¥n is true in an
¤ ¤
1 i j n
L -structure iff the structure has at least n elements in its domain.
Now consider Γ t φu Y tD¥n : n ¥ 1u.
Our first subclaim is that Γ is unsatisfiable, because φ is satisfiable in
all and only finite structures, and the set tD¥n : n ¥ 1u is satisfiable in all
and only infinite structures.
The second subclaim is that any finite subset of Γ is satisfiable. You
should convince yourself of this by thinking about what a finite subset of
tD¥n : n ¥ 1u looks like.
Putting the two subclaims together shows that Γ contradicts the com-
pactness of L . Thus there is no such formula φ.
20
Lecture 5
The first half of the course consisted of semantic metatheory, principally that
of L1 . We were concerned with the semantic consequence relation (, and
with proving general results about the logic, of the form say ‘If Γ ( φ then...’,
rather than with proving specific results such as say tP, P Ñ Qu ( Q. In the
second half, we’ll focus more on deductive metatheory. In this lecture, we’ll
think about deductive systems in a more abstract way than you’ve hitherto
been used to. In later lectures the focus will be on the specific proof system
you studied in Introduction to Logic, especially its propositional fragment.
A logic L may be characterised in general terms as consisting of a language
and a consequence relation (L . SenpLq is the set of sentences of the logic,
and the consequence relation (L is almost always taken to relate subsets
of SenpLq to elements of SenpLq; that is, if Γ (L φ then Γ SenpLq
and φ P SenpLq. (Formally speaking, then, the relation (L is a subset of
PpSenpLqq SenpLq; don’t worry if this way of putting things doesn’t make
sense.) The relation (L is not usually taken as primitive but rather defined
using the notion of an L-structure, as in the Introduction to Logic course,:
Γ (L φ just when, for all L-structures A, if A satisfies all the elements of Γ
then A satisfies φ.
A deductive system D for a logic gives rise to a consequence relation $D ,
where Γ $D φ is usually taken to mean: there’s a proof in D of φ from
premises all of which are elements of Γ. It’s tricky to say exactly what we
require of a deductive system and what it means to be a proof. Part of the
job description of a deductive system is that it be syntactic, i.e. concerned
only with symbols and not their meanings; but spelling this out is not as
straightforward as you might think. Proofs are also required to be decidable:
there should be an algorithm (a computer programme if you like) that returns
the answer YES when faced with a string of symbols that is a D-proof and
NO when it isn’t. A string of symbols may not be a D-proof either because
it’s not a sequence of L-formulas, or because it is but the formulas are not
combined together in the right way to make up a D-proof. We’ll set to one
side all these difficult questions about what a proof system is and just assume
that it satisfies three conditions.
21
iii. For all Γ, ∆ SenpLq and all φ P SenpLq, if Γ $D φ then Γ Y ∆ $D φ.
We used the first condition at the end of lecture 4 to prove the compact-
ness theorem. The condition states that proofs can only have finitely many
premises. The second condition expresses the intuitive thought that one can
prove anything one assumes. And the third condition expresses the idea that
provability is monotonic: unlike inductive reasoning, in deductive reasoning
the set of conclusions you can prove can’t shrink by adding more premises.
These three requirements are fairly weak conditions to put on a proof sys-
tem. I certainly wouldn’t want to claim that satisfying them is sufficient for
being a proof system. In fact, the conditions may not even be necessary: per-
haps something deserves to be called a deductive system even whilst failing
one or more of the conditions.
Moving on, let’s consider how (L and $D may be related.
I’ve put the soundness condition in brackets in the definitions of strong and
weak completability because it’s usually assumed that any proof system we
might be interested in is sound. Note that $D φ means that φ is provable in
D from the empty set, so could be alternatively written as H $D φ; similarly
(L φ means that φ is semantically entailed from the empty set and could
be written as H (L φ. It’s immediate, then, that the strong versions of the
theses imply the weak versions. Logicians often omit the adjectives ‘weak’
and ‘strong’, it being clear from context which they mean.
In Introduction to Logic, you saw that the ND system is sound and com-
plete with respect to L -consequence. It will be useful to have labels for the
three systems you encountered there.
Definition 16 Let N Di be the system of rules in The Logic Manual for the
Li -connectives, for i 1, 2, .
22
Thus N D1 consists of all and only the propositional rules, N D2 extends N D1
with the rules for @ and D, and N D in turn extends N D2 with the rules
for . As was mentioned in that course, N Di is sound and complete with
respect to Li -consequence, for i 1, 2 and . Caveat: it does not in general
follow from the fact that logic L with deductive system D is a sublogic of L
with deductive system D that if Γ $D φ for Γ SenpLq and φ P SenpLq
then Γ $D φ. When no more L-sequents are proved by D than by D, we
say that the former system is conservative over the latter. In more advanced
proof theory, the notion of conservativeness will turn out to be of central
importance.
Let’s now do some elementary ‘abstract proof theory’. In what follows,
we assume that SenpLq is non-empty and that it is closed under negation,
i.e. if φ P SenpLq then φ P SenpLq. We use Γ &D φ to mean that it’s not
the case that Γ $D φ.
How are these two notions related? ConsistencyD is more general than
negation-consistencyD , because it does not assume the existence of negation
in the language. It’s also weaker, even granted that assumption. To spell all
this out, let’s see first why negation-consistencyD implies consistencyD . It’s
understood throughout that Γ SenpLq and φ P SenpLq.
23
Proof Suppose Γ is negation-consistentD . We saw in the previous lemma
that Γ is consistentD . Γ also vacuously satisfies the condition for underwriting
EFQ since there is no φ such that Γ $D φ and Γ $D φ (if the antecedent
of a conditional begins ‘there exists an F such that ...’ then the conditional
is true when there is no such F ).
Suppose conversely that Γ is consistentD and that D underwrites EFQ
from Γ. If Γ were negation-inconsistentD , there would be a φ such that
Γ $D φ and Γ $D φ. Because D underwrites EFQ from Γ, it would follow
that Γ $ ψ for all ψ. Hence Γ would be inconsistentD , a contradiction.
The global conditions DNI, DNE and RED are, as you may have guessed,
overkill for these results.
24
You should check that N D1 underwrites DNI, DNE and RED, as well
as EFQ. Going forward, we’ll assume all these facts about N D1 , e.g. we’ll
assume that Γ Ytφu is consistentN D1 iff Γ &N D1 φ, for all Γ SenpL1 q and
φ P SenpL1 q.
25
Lecture 6
Let’s continue to think about deductive systems abstractly before focusing
on a concrete proof sytem, that of N D1 (the propositional fragment of the
proof system in The Logic Manual ). As usual, we assume that Γ is a set of
sentences and φ a sentence, and in the following definitions we also assume
that our logic L is closed under negation (if φ is in SenpLq then so is φ).
26
We note that the lemma’s converse fails. Consider for example Γ tα : α
is a sentence letteru. You should check that Γ is N D1 -consistent and N D1 -
complete. Yet Γ is patently not maximally N D1 -consistent; e.g. Γ $N D1
P ^ Q, so Γ Y tP ^ Qu is N D1 -consistent yet P ^ Q R Γ.
The next lemma tell us that for maximally N D1 -consistent sets, mem-
bership coincides with derivability.
Next, we show that consistency and completeness are sufficient for deriv-
ability to behave just like truth in a structure.
Proof I’ll prove the first two parts and leave the rest to you. We assume
thoughout that Γ is N D1 -consistent and N D1 -complete. We’ll also assume
the equivalence of consistencyN D1 and negation-inconsistencyN D1 , previously
proved.
For the left-to-right direction in (i), suppose Γ $N D1 φ . Since Γ is
N D1 -consistent, it follows that Γ & φ.
For the right-to-left-direction in (i), suppose Γ & φ. Then by Γ’s N D1 -
completeness, Γ $ φ.
27
For the right-to-left direction in (ii), suppose Γ $N D1 φ and Γ $N D1 ψ.
Let Π1 be an N D1 -proof whose set of undischarged assumptions is a subset of
Γ and whose conclusion is φ, and similarly let Π2 be an N D1 -proof whose set
of undischarged assumptions is a subset of Γ and whose conclusion is ψ. We
let Π3 be the proof obtained by putting Π1 and Π2 side by side and applying
the ^-introduction rule to their respective conclusions φ and ψ to obtain
φ ^ ψ. Π3 is thus an N D1 -proof whose set of undischarged assumptions is
a subset of Γ (since the sets of undischarged assumptions of Π1 and Π2 are
subsets of Γ) and whose conclusion is φ ^ ψ. Pictorially:
Π1 Π2
.. ..
. .
φ
φ^ψ
ψ
^intro
For the left-to-right direction in (ii), suppose Γ $N D1 φ ^ ψ. A similar
argument to the one just given shows that a proof of φ ^ ψ from premises that
are all elements of Γ can be extended by a single application of the first ^-
elimination rule to a proof of φ from these same premises, and that this same
proof can be extended by a single application of the other ^-elimination rule
to a proof of ψ from these same premises. Hence Γ $N D1 φ and Γ $N D1 ψ.
Cases (iii), (iv) and (v) are entirely analogous.
(i) φ P Γ iff φ R Γ
Proof A consequence of the previous two lemmas. We’ll do cases (i) and
(ii). We assume that Γ is maximally N D1 -consistent throughout.
As for (i): by Lemma 26, φ P Γ iff Γ $N D1 φ. By Lemma 27, Γ $N D1
φ iff Γ &N D1 φ. And by Lemma 26 again, Γ &N D1 φ iff φ R Γ. So φ P Γ
iff φ R Γ.
28
As for (ii): by Lemma 26, φ ^ ψ P Γ iff Γ $N D1 φ ^ ψ. By Lemma 27,
Γ $N D1 φ ^ ψ iff Γ $ ψ and Γ $ φ. And by Lemma 26 again, Γ $ φ and
Γ $ ψ iff φ P Γ and ψ P Γ . So φ ^ ψ P Γ iff φ P Γ and ψ P Γ.
Cases (iii), (iv) and (v) are entirely analogous.
29
Lecture 7
Last time, we familiarised ourselves with maximally consistent sets and their
properties, as well as the properties of complete and consistent sets, of which
maximally consistent sets are an example. Today, we’ll prove the soundess
of N D1 with respect to L1 -consequence, and make a start on proving its
completeness. Without further ado, then, let’s prove soundness.
_ Q _ intro
P
1
P
pP _ Qq _ R _ intro 1
has size 2 since it applies the (first) _-introduction rule twice. We then
prove the result by induction on the size of proofs, taking as our inductive
hypothesis that if Π is an N D1 -proof of φ from Γ of size N then Γ ( φ.
For the induction basis, suppose Π is a proof of size 0, i.e. containing no
rule applications. Then Π must take the form: φ. Since this one-line proof
is a proof from Γ of φ, φ must be an element of Γ. Clearly, then, Γ ( φ as
φ P Γ.
For the inductive step, assume the inductive hypothesis for proofs of size
¤ N . So let Π be a proof of size N 1 of φ from a set of undischarged
assumptions all of which are elements of Γ. Let’s call the last rule used
in this proof ρ (notice that every proof has a last rule). The argument
now proceeds by considering all the possibilities for ρ. This involves a large
number of cases, so I’ll do one here and leave the rest to you.
Suppose ρ is the ^-introduction rule. So Π takes the following form:
Π1 Π2
.. ..
. .
ρ ^intro
φ1 φ2
φ1 ^ φ2
30
where φ φ1 ^ φ2 . Let’s call P rempΠi q the set of undischarged assumptions
in subproof Πi of φi , for i 1, 2. Since P rempΠ1 q and P rempΠ2 q are both
subsets of the set of undischarged assumptions in Π, which itself is a subset
of Γ, it follows that P rempΠ1 q, P rempΠ2 q Γ. And since Π1 and Π2 are of
size ¤ N , it follows from the inductive hypothesis that
By the semantic rule for conjunction, we deduce that P rempΠ1 qYP rempΠ2 q (
φ1 ^ φ2 . And since P rempΠ1 q Y P rempΠ2 q Γ, we futher deduce
Γ ( φ1 ^ φ2
This proves the inductive step for the case ρ ^-introduction. I leave the
cases in which ρ is some other rule as exercises for you, which you should
endeavour to do. They are very similar to the above. (A detail that’s un-
likely to trip you up yet that’s nevertheless worth mentioning: for rules that
discharge assumptions you must be careful not to assume that Π’s set of
undischarged assumptions is the union of the set of undischarged assump-
tions of its subproofs.)
31
It’s immediate from this definition that Γn is N D1 -consistent for all n. We
can prove this by induction: Γ0 Γ is N D1 -consistent by assumption, and
Γn 1 is clearly N D1 -consistent if Γn is.
Let’s now
define Γ as the set of sentences that appear in any of the
Γn : Γ Γn . Clearly, Γn Γ for all n, including the case n 0, i.e.
¤
0 n
Γ Γ0 Γ . We first prove that Γ is N D1 -consistent before proving that
it’s maximally N D1 -consistent.
Suppose for reductio, then, that Γ were N D1 -inconsistent, so that Γ $N D1
φ and Γ $N D1 φ. (As ever, we’re taking negation-consistencyN D1 and
consistencyN D1 as equivalent.) Thus, since proofs are finite:
The First Auxiliary Lemma (Lemma 30) is the first staging post on the
way to proving N D1 -completeness. We now state and prove the second.
32
Proof Assume Γ is maximally N D1 -consistent. We define AΓ so that for
atomic formulas α (i.e. sentence letters),
AΓ ( α iff α P Γ
We must now prove that this applies to all formulas φ, i.e. that φ P Γ iff
AΓ ( φ whatever φ may be. The proof is by induction on the complexity of
φ.
Base case: φ is of complexity 0, i.e. φ is a sentence letter. There is
nothing to prove, since φ P Γ iff AΓ ( φ holds by the definition of AΓ .
For the induction step, suppose φ is of complexity N 1. There are, as
you would expect, five cases to consider.
The first case is φ ψ, where ψ has complexity N . Since Γ is maximally
N D1 -consistent, then by Corollary 28, ψ P Γ iff ψ R Γ. By the induction
hypothesis, ψ P Γ iff AΓ ( ψ, or equivalently, ψ R Γ iff AΓ * ψ. And
clearly, by the semantic rule for , AΓ * ψ iff AΓ ( ψ. From these three
biconditionals, we deduce that ψ P Γ iff AΓ ( ψ.
The second case is φ φ1 ^ φ2 , where φ1 and φ2 each has complexity
¤ N . Since Γ is maximally N D1-consistent, then by Corollary 28, φ1 ^ φ2 P Γ
iff φ1 P Γ and φ2 P Γ. By the induction hypothesis, φi P Γ iff AΓ ( φi , for
i 1, 2. And clearly, by the semantic rule for ^, AΓ ( φ1 ^ φ2 iff AΓ ( φ1
and AΓ ( φ2 . From these three biconditionals, we deduce that φ1 ^ φ2 P Γ
iff AΓ ( φ1 ^ φ2 .
The third, fourth and fifth cases, dealing with _, Ñ and Ø are entirely
analogous and are left as exercises.
33
Lecture 8
Last time, we proved the soundness of N D1 with respect to L1 ’s consequence
relation. We also proved that any N D1 -consistent set can be extended to a
maximally consistent set (First Auxiliary Lemma, i.e. Lemma 30) and that
a maximally consistent set corresponds to a unique L1 -structure (Second
Auxiliary Lemma, i.e. Lemma 34). It now remains to put these pieces
together to prove N D1 ’s completeness with respect to L1 -consequence.
The proof is deceptively simple, because most of the work has been packed
into the First Auxiliary Lemma and the Second Auxiliary Lemma. Moreover,
the Second Auxiliary Lemma itself depended on a host of lemmas about the
properties of maximally N D1 -consistent sets.
The proof we gave at the end of Lecture 4 of the Compactness Theorem
made use of Soundness and Completeness. Since we’ve now proved these for
N D1 , we’ve discharged our obligations as far as L1 is concerned: we’ve proved
that (L1 is compact. But as I mentioned in lectures, it would be better if
possible to give a purely semantic proof of that semantic result than have to
give a deductive argument. In the last part of today’s lecture, we’ll sketch
a strictly semantic proof of L1 ’s compactness. The proof will effectively be
a semantic version of the argument for N D1 ’s completeness. So it will also
serve the purpose of highlighting the proof’s more abstract features.
Recall Definition 12, which stated that a set of sentences is finitely sat-
isfiable just when all its finite subsets are satisfiable. L1 ’s compactness is
equivalent to the statement that if a set of sentences Γ is finitely satisfiable
then it’s satisfiable – this was left as an exercise back in lecture 4, which I
urge you to do. As we’re going to try to mimic the proof of N D1 ’s complete-
ness as much as possible in proving the compactness of L1 , we’ll need the
following definition.
34
Definition 21 Γ is maximally finitely satisfiable just when Γ is finitely sat-
isfiable and if Γ Y tφu is finitely satisfiable then φ P Γ.
I’ll give abbreviated proofs of each of these lemmas, since their proofs are
so similar to that of their deductive counterparts.
35
Given the fact that φ P Γ iff φ R Γ, it’s now easy to prove that (i)
φ1 ^ φ2 P Γ iff φ1 P Γ and φ2 P Γ; (ii) φ1 _ φ2 P Γ iff φ1 P Γ or φ2 P Γ (or
both); (iii) φ1 Ñ φ2 P Γ iff φ1 R Γ or φ2 P Γ (or both); and (iv) φ1 Ø φ2 P Γ
iff (φ1 P Γ and φ2 P Γ) or (φ1 R Γ and φ2 R Γ). For example, if φ1 ^ φ2 P Γ and
φ1 R Γ then tφ1 ^ φ2 , φ1 u would be a finite but unsatisfiable subset of Γ. We
conclude that membership in Γ behaves just like truth in an L1 -structure.
Using the above, we define AΓ as in the proof of the Second Auxiliary
Lemma, so that for atomic formulas α (i.e. sentence letters),
AΓ ( α iff α P Γ
It’s now easy to prove that φ P Γ iff AΓ ( φ for all all formulas φ (not just
atomic ones). The proof is once more by induction on the complexity of φ,
using the facts established in the previous paragraph.
Combining the First Auxiliary Lemma* and the Second Auxiliary Lemma*
yields an alternative proof of the compactness of L1 (an instance of theo-
rem 19).
You may have been wondering about how our proofs of N D1 -completeness
and L1 ’s compactness would go if the set of sentence letters were not count-
ably infinite. Clearly, if there were only finitely many sentence letters the
proofs would be easier if anything; an alternative proof of compactness could
be given from the observation that there’s a finite subset of sentences such
that every sentence is logically equivalent to one of its members. So the
question is what happens in the case in which the set of sentence letters is
uncountable. In fact, one can give a more abstract version and (once one
has got used to its initally dizzying level of abstraction) easier version of the
arguments for the First Auxiliary Lemma and the First Auxiliary Lemma*.
Here we’ll give the argument for the latter lemma, easily amended to yield
the argument for the former lemma. (The next paragraph is non-examinable,
and uses some terminology about orders that I won’t define but invite you
to look up.)
36
Suppose Γ is finitely satisfiable. Order by inclusion the set FΓ of finitely-
satisfiable sets of sentences of the language containing Γ. FΓ is non-empty,
since it contains at least Γ. Any chain in FΓ has an upper bound, obtained
by taking the union of the elements in the chain: this union contains Γ as a
subset since all the members of the chain do, and it is finitely satisfiable since
any of its finite subsets must come from some element of the chain, which by
hypothesis is finitely satisfiable. Zorn’s Lemma states precisely that every
partial order with the property that every chain has an upper bound has a
maximal element. Since the conditions of Zorn’s Lemma are satisfied, we
deduce that FΓ has a maximal element; that is, FΓ is a maximal finitely
satisfiable set extending Γ. Note that nowhere did we rely on the fact that
the sentence letters are denumerably many, or on any assumption about the
set of connectives. So this more general argument establishes the analogue
of the First Auxiliary Lemma* for a propositional logic with atom set of any
size.3
Finally, I leave you with a teaser. As we’ve seen, the three logics studied
in Intro Logic are compact. A natural question is whether English itself is
compact. To tackle this question, one must first clarify what compactness
means for a non-formal language such as English. A reasonable definition
might be that if an English argument Γ 6 δ is valid, where Γ is a set of
English sentences and δ is an English sentence, then Γf in 6 δ is valid, where
Γf in is a finite subset of Γ. What validity in English comes to is of course
a vexed issue. But put sceptical doubts aside for a minute and assume that
this notion is in good order. Consider now the following argument:
37
no finite subset of the premiss set entails the conclusion; that there are at
least n1 , nk things for some finite n1 , nk does not entail that there are
infinitely many things.
If this is right—and I certainly haven’t proved it—then English conse-
quence is incompact, and none of L1 , L2 and L is capable of capturing it,
since these three logics are all compact. Indeed, as we saw in lecture 4, the
strongest of the three, L , isn’t even capable of formulating the argument’s
conclusion. So which logic captures English validity?
38