A Metalogical Primer: 1 Notation
A Metalogical Primer: 1 Notation
A Metalogical Primer: 1 Notation
Jason Turner
Spring 2010
Metalogic is the study of logic. When you learn a logic, you learn a logical
system: a system of rules for taking sentences of a given language (usually, a
symbolic language) and proving other sentences using those rules.
After you’ve messed about with logical systems for a bit, you can start asking
questions about those systems. Like: can I construct a proof for every valid
argument, or only for some of them? Or: are there any valid arguments with
infinitely many premises, where each of the premises is needed? (That is, where
if you took any finite subset of the premises, the conclusion wouldn’t follow
from it?) These questions are the questions of metalogic.
1 Notation
When we do proofs, we use expressions (of our formal, technical language).
But when we do metalogic, we want to talk about expressions of this language,
and the way expressions of this language are related to each other. To avoid
confusion — to make it clear whether we’re using a sentence, or talking about
it — we’ll use a certain kind of notation.
First of all, if we want to talk about particular expressions, we use quotation-
names to talk about them. For instance, if I want to say that the third letter of
the alphabet is curved, I would say that ‘C’ is curved. See how I used the quotes
there? This is important. In the following sentences,
(2) is true — ‘Leeds’ is a word, and since it begins with ‘L’, it’s in the middle of
the dictionary. But Leeds is not a word — it’s a city — and cities can’t fit in the
middle of dictionaries.
So far, so good. But what if we don’t want to talk about a particular word or
sentence, but a whole bunch of words or sentences of thus-and-so a type? That
is, what if we want to quantify over expressions?
In this case, to avoid confusion, we adopt a convention. Different authors
will use different conventions. The convention I’ll use for this document is
Quine’s corner-quote convention. In this convention, we use greek characters as
‘metalinguistic’ variables — variables that go as proxy for expressions. Using
this convention we can say, for instance, that every word φ that begins with ‘L’
is closer to the middle of the dictionary than every word χ that begins with ’B’.
Sometimes, however, we want to talk about a whole bunch of complex ex-
pressions that all have the same form. For instance, we might want to talk about
all sentences that have the following form: their first part is a sentence, and then
1
the word ‘and’ occurs, and then their final part is another sentence. Talking like
this is a bit of a mouthful. Quine’s corner-quote convention lets us instead use
corner-quotes. In this convention, I can say something about all sentences of the
form pφ and ψq. The expression
pφ and ψq
the result of writing φ and then leaving a space and then writing
‘and’ and then leaving another space and then writing ψq
But again, that’s a mouthful to say, so it’s quicker and easier to use the corner-
quote notation instead.
2
Second, the formulation rules. These form a recursive definition of ‘sentence’:
they tell us what the basic sentences are, and how to build up more complex
sentences from that. The rules:
Notice that, technically speaking, ‘A & B’ is not a sentence, but ‘(A & B)’ is.
There are good reasons for this (they have to do with making sure the defini-
tions don’t screw up scope distinctions), but they create clutter because every
compound sentence has an outermost pair of brackets. So we adopt the conven-
tion of not writing the outermost pair of brackets, while secretly acknowledging
that, strictly speaking, this is a mistake.
What about the semantics — the meaning of these expressions? Well, we tend
to be silent about just what the atomic propositions mean, because it turns out it
doesn’t matter what they mean to the study of logic. All we demand is that they
have meanings which can be either true or false. But we do specify meanings
for the sentential connectives, as follows:
Connective Meaning
∼ not
∨ or
& and
⊃ if,. . . then
≡ if and only if
3
However, the reason we use that procedure and call it a ‘proof’ procedure is
because we know that ‘&’ means ‘and’. If ‘&’ had meant ‘or’ instead, although
this procedure would still exist, we wouldn’t want to use it or call things made
from it ‘proofs’. And we wouldn’t teach it to computers, and so on.
There are, in general, three kinds of proof procedures that get used: natural
deduction, semantic tableaux (or ‘truth-trees’), and axiomatic systems. In your
introduction to logic course, you probably learned one of the first two. In the
first sort, you have a whole lot of different rules, including ones involving mak-
ing and discharging assumptions, that let you ‘infer’ one sentence from some
others. In the second sort of procedure, you have rules for constructing trees
from an initial list of sentences, and then ‘closing off’ ones with contradictions
on them.
But we will focus on the third sort here, because it ends up being more useful
for doing metalogic. An axiom system consists of
(A1) φ ⊃ (ψ ⊃ φ)
(A2) (φ ⊃ (ψ ⊃ χ)) ⊃ ((φ ⊃ ψ) ⊃ (φ ⊃ χ))
(A3) (∼φ ⊃ ∼ψ) ⊃ ((∼φ ⊃ ψ) ⊃ φ)
(D≡) (φ ≡ ψ) ⊃ ∼((φ ⊃ ψ) ⊃ ∼(ψ ⊃ φ))
(D&) (φ & χ) ≡ ∼(φ ⊃ ∼χ)
(D∨) (φ ∨ χ) ≡ (∼φ ⊃ χ)
Notice that what gets put in for φ, ψ, and χ in the axioms can be simple or
complex:
A ⊃ (B ⊃ A), and
(P ≡ ∼R) ⊃ (((∼B ∨ C) ⊃ ∼R) ⊃ (P ≡ ∼R))
are equally good instances of (A1), and are equally well axioms. The difference
is that, in the second one, ‘P ≡ ∼R’ is put in for φ (note how I remembered to
add back the outermost brackets!) and ‘(∼B ∨ C) ⊃ ∼R’ is put in for ψ.
Let’s work through a simple axiomatic proof: we want to prove, from the
premise ‘∼P ⊃ P’, the conclusion P. We begin our list with the premise:
4
(1) ∼P ⊃ P
Now, putting ‘∼P’ in for φ and ‘∼P ⊃ ∼P’ in for ψ, we get an instance of (A1),
which we’ll make the second entry on our list:
This time, we’ll substitute ‘∼P’ in for both φ and χ in (A2), along with ‘∼P ⊃
∼P’ for ψ:
(3) (∼P ⊃ ((∼P ⊃ ∼P) ⊃ ∼P)) ⊃
((∼P ⊃ (∼P ⊃ ∼P)) ⊃ (∼P ⊃ ∼P))
Since the antecedent of (3) just is sentence (2), we can use modus ponens to get:
Now we substitute ‘∼P’ for both φ and ψ into (A1) to get a different axiom
instance:
(6) ∼P ⊃ ∼P
(8) (∼P ⊃ P) ⊃ P
Now we get
(9) P
(1) ∼P ⊃ P Premise
(2) ∼P ⊃ ((∼P ⊃ ∼P) ⊃ ∼P) A1
(3) (∼P ⊃ ((∼P ⊃ ∼P) ⊃ ∼P)) ⊃
((∼P ⊃ (∼P ⊃ ∼P)) ⊃ (∼P ⊃ ∼P)) A2
(4) (∼P ⊃ (∼P ⊃ ∼P)) ⊃ (∼P ⊃ ∼P) MP: 2, 3
(5) ∼P ⊃ (∼P ⊃ ∼P) A1
(6) ∼P ⊃ ∼P MP: 4, 5
(7) (∼P ⊃ ∼P) ⊃ ((∼P ⊃ P) ⊃ P) A3
5
(8) (∼P ⊃ P) ⊃ P MP: 6, 7
(9) P MP: 1, 8
As you can see, this a list where every line is either (i) a premise (line 1), (ii)
an instance of an axiom (lines 2, 3, 5, and 7), or (iii) licensed by modus ponens
from earlier lines (4, 6, 8, and 9).
Proving the same result from the same premises would have been a lot easier
in either natural deduction or semantic tableaux systems. In general, proofs
are more difficult to construct in an axiomatic system than in either natural
deduction or semantic tablueaux systems.
So why do we use them? Because, although actually doing logic is harder
in them, reasoning about logic is easier. And, as it turns out, whatever can be
proven in an axiomatic system can be proven in a natural deduction or tableaux
system, or vice versa. So we can do our metalogic focusing (mainly) on axiomatic
systems, secure in the knowledge that our results will apply equally well to
other types of proof procedures.
If ∆ is a set of sentences, and φ is a sentence, then we say that ∆ proves φ, or
∆ φ, if and only if there is a (possible) proof of φ with only sentences in ∆ as
premises. If ∅ φ — that is, if φ can be proved with a proof where every line is
either an axiom or comes in by modus ponens — then we call φ a theorem, and
write φ.
6
vi) If φ is a biconditional of sentences ψ and χ — that is, if φ =
pψ ≡ χq — then φ is true on V iff ψ and χ are either both true
on V or both not true on V.
(∗) ∼A ⊃ (B ∨ A)
A B ∼A ⊃ (B ∨ A)
T T T
T F T
F T T
F F F
Each row on the table represents a different truth-value assignment — that is, a
different model. Or, more accurately, each row represents a class of truth-value
assignments. The first row, for instance, represents all the assignments V where
V (A) = 1 and V (B) = 1; the second represents all V where V (A) = 1 and
V (B) = 0; and so on.
The entries under each sentence tell us whether that sentence is true or false
on the model (or, more appropriately, on all the models) associated with the
given rows. For instance, this truth-table tells us that, on all models V where
1 By ‘false on V’, I just mean ‘not true on V’.
7
V (A) = 0 and V (B) = 0, the sentence (∗) is false, and that on all other models,
(∗) is true. Truth-tables are thus an easy way for verifying facts about all mod-
els. Something like this might have been emphasized in your introduction to
logic course. What probably wasn’t emphasized was that the truth-table method
works because truth-on-a-model is defined via clauses (i)–(vi) above.
Finally, some definitions. What we mainly want to define is model-theoretic
entailment: the relation that some sentences hold to another sentence just in
case the first can’t be true on a model without the second also being true on that
model. We can define it ‘officially’ as follows. First, if ∆ is a set of sentences,
say that ∆ is true on a model iff all of its members are true on that model. Then
we say that ∆ entails φ if and only if every model V on which ∆ is true is
also one on which φ is true. We write this ∆ φ. If ∅ φ, we say that φ is a
model-theoretic validity.
If ∆ φ, then ∆ φ
If ∆ φ, then ∆ φ
8
As it turns out, the propositional calculus — the system of propositional
logic we have been looking at — is both sound and complete.
2.4.1 Soundness
First, here’s how we prove soundness. Suppose ∆ ψ. What we want to show
is that, ∆ φ. But to say that ∆ φ is just to say that, for every model M, if all
the sentences in ∆ are true on M, then φ is true on M, too.
So what we do is we begin by assuming two things:
ψ1
ψ2
..
.
ψn
9
But remember that we are assuming that all the members of ∆ are true on M
(assumption (ii)), so if it got in that way, we know it’s true on M.
But what if ψ1 is an axiom? Well, here’s where some of the important bits
come in. We want to show that, if ψ1 is an axiom, it is true on M. But we don’t
know which axiom it is, and we don’t know which model M is, either. So what
we need to show is that all axioms are true on all models. If that’s the case, then
of course ψ1 is true on M — if all axioms are true on all models, then of course
this axiom is true on this model.
How do we do this? Well, a sentence is true on all models if and only if it
is a model-theoretic validity. And it’s a model-theoretic validity if and only if
it shows up as ‘T’ on every line of its truth-table. So we can check each axiom
schema to see if it will have ‘T’s on every line of its truth-table. For instance, the
first axiom schema comes to:
φ ψ φ ⊃ (ψ ⊃ φ)
T T T T
T F T T
F T T F
F F T T
So any instance of the first axiom schema is a model-theoretic validity, and so
is true on all models. You can use truth-tables to check that the same holds for
the other axiom schemas; I won’t do that here. But then all axioms are true on
all models, so — if ψ1 is an axiom, then ψ1 is true on all models too, and so is
true on M.
So we know that ψ1 is true on M. Now we need to show that, for any sen-
tence ψi in the sequence, if all the sentences that come before it in the sequence
are true on M, then it is, too. So suppose that all the sentences that come before
ψi are true on M. Is ψi also true on M?
Well, as above, ψi either got in the sequence by (1) being a member of ∆, or
by (2) being an axiom, or by (3) modus ponens. We’ve already seen why it will
be true on M if it got in in by (1) or (2); that reasoning still applies. So we only
need to make sure that it will be true on M if it got in by modus ponens.
So: assume that all the sentences before ψi are true on M, and ψi got in by
modus ponens. Then there must be some sentence ψj such that the sentences
ψj
ψj ⊃ ψi
are both in the sequence somewhere before ψi . So they must both be true on M.
But by our definition of truth-on-M, if pψj ⊃ ψi q is true on M, then either ψj is
false on M or φi is true on M. But we know that φj is not false on M — it’s true
on M — so we know that ψi must be true on M, by our definition of truth on a
model.
But this completes our proof: We now know that every sentence of the proof,
no matter how it got into the proof, is true on M, which means that φ — the
final sentence of the proof — is true on M, too.
10
2.4.2 Completeness
Completeness:
If ∆ φ, then ∆ φ
is thought harder to prove than soundness. What we do is to focus not on
completeness as it’s usually stated, but on its contrapositive:
If ∆ 6 φ, then ∆ 6 φ
Here’s how we do it. Say that a set of sentences Γ is consistent if you cannot
use it to prove a contradiction — that is, if Γ 6 φ & ∼φ. Then we prove
completeness by proving the following three lemmas:
Lemma 1: If ∆ 6 φ, then ∆ ∪ {∼φ} is consistent.
11
it can. That is, if any atomic sentence is not in S, then that atomic sentence is
inconsistent with the members of S.
So let’s let M be a model where, for any atomic sentence χ, M(χ) = 1 if and
only if χ ∈ S. Now we need to show that, if ψ ∈ S, then ψ is true on M. We
also do this by a mathematical induction, this time an induction on how many
connectives ψ has in it.
More precisely, we show that, for all sentences ψ, ψ ∈ S iff ψ is true on M.
We start by showing that this holds for cases where ψ has zero connectives —
that is, where ψ is atomic. (This should be easy — it just comes out of how we
constructed M from S.) Then we show that, if all sentences shorter that ψ have
this property, then ψ does to. The proof starts by noticing that ψ will have the
form ψi Cψ2 for some connective C. And ψ1 and ψ2 will be true on M iff they
are members of S. So now all we need to show is that ψ is true on M iff it is a
member of S, too.
We do this by going case-by-case through all the connectives that C might
be. I’ll just do one, to show you how it goes.
Suppose that C is the conjunction connective, ‘&’. We know that ψ1 ∈ S iff
ψ1 is true on M, and same for ψ2 .
Suppose first (for reductio) that ψ ∈ S but ψ is not true on M. Since ψ is a
conjunction of ψ1 and ψ2 , it is consistent with both of these; since it’s consistent
with both of these, and is in S, then both ψ1 and ψ2 must be in S, too. But if
they’re both in S, since they have the property we care about, they must both be
true on M — in which case ψ will be true on M, too. Contradition.
Suppose second (for reductio) that ψ < S but ψ is true on M. Since ψ is true
on M, and is the conjunction of ψ1 and ψ2 , both of these must be true on M,
too. But if they’re both true on M, then since they have the property we care
about, they must both be in S. But ψ1 and ψ2 are clearly consistent with their
conjunction, so, since S is maximal, their conjunction, which is ψ, must be in S,
too. Contradiction.
We’ve looked at the only two ways ψ could fail to have the property; since
they both lead to contradiction, ψ must have the property after all.
12