The Content of Deduction
Mark Jago
Forthcoming in Journal of Philosophical Logic. Draft of November 2011.
Abstract: For deductive reasoning to be justified, it must be guaranteed to preserve truth
from premises to conclusion; and for it to be useful to us, it must be capable of informing
us of something. How can we capture this notion of information content, whilst respecting
the fact that the content of the premises, if true, already secures the truth of the conclusion?
This is the problem I address here. I begin by considering and rejecting several accounts
of informational content. I then develop an account on which informational contents are
indeterminate in their membership. This allows there to be cases in which it is indeterminate
whether a given deduction is informative. Nevertheless, on the picture I present, there are
determinate cases of informative (and determinate cases of uninformative) inferences. I
argue that the model I offer is the best way for an account of content to respect the meaning
of the logical constants and the inference rules associated with them without collapsing
into a classical picture of content, unable to account for informative deductive inferences.
Keywords: Content, information, deduction, inference, epistemic scenarios.
1 Introduction
D
eductive reasoning is essential to philosophy, mathematics and logic.
What is not so clear is how deductive reasoning conveys information to
us. In ‘The Justification of Deduction’, Dummett (1978, 297) asks how
deduction can be both justified and useful. If it is justified, it must be guaranteed
to preserve truth from premises to conclusion. To be useful, it must inform us
of something.1 But how, wonders Dummett, can the move from premises to
conclusion be informative, if the former already guarantee the latter?2 The task is
to capture this notion of information content whilst respecting the fact that the
content of the premises, if true, already secures the truth of the conclusion. This is
the aim of this paper.
According to a popular analysis, for a proposition to be informative is for it to
rule out certain scenarios, or would-be possibilities.3 The proposition that it rarely
rains in Cambridge is informative because it excludes possible scenarios in which
it rains regularly in Cambridge. Before coming to believe that proposition, it was
possible as far as I was concerned that it rains regularly in Cambridge. In coming
to believe that proposition, I cease to treat such scenarios as ways the world might
1. ‘For [deduction] to be legitimate, the process of recognising the premisses as true must already
have accomplished whatever is needed for the recognition of the truth of the conclusion; for it to be
useful, a recognition of its truth need not actually have been accorded to the conclusion when it was
accorded to the premisses’ (Dummett 1978, 297).
2. As Dummett notes, ‘no definite contradiction stands in the way of satisfying these two requirements’;
yet ‘it is a delicate matter so to describe the connection between premisses and conclusion as to display
clearly the way in which both requirements are fulfilled’ (Dummett 1978, 297).
3. Hintikka (1962) gives the classic presentation of the view; Lewis (1975; 1986), Stalnaker (1976;
1984) and Chalmers (2002; 2010) put it to work in various ways. For recent overviews of the literature,
see van Benthem and Martinez 2008 and van Benthem 2011.
1
be, for all I know. We can think of all scenarios according to which it rarely rains
in Cambridge as constituting a notion of content for ‘it rarely rains in Cambridge’
which is suitable for various epistemic purposes.4 To believe that proposition is to
treat no other scenario as being a doxastic possibility; to know that proposition
is to treat no other scenario as being an epistemic possibility. That content is
informative for an agent iff coming to believe (or know) that proposition narrows
down her doxastically (or epistemically) accessible scenarios. To be informative at
all, therefore, a statement must have a non-empty content.
How should we think of the content of a valid deduction Γ ⊢ A, from premises
Γ to conclusion A, within this framework?5 We can think in terms of the differences
in an agent’s belief state in the move from believing each premise in Γ but not
believing the conclusion A to believing A as well as all premises in Γ. Alternatively,
we can think in terms of an agent discovering the incompatibility of the premises
Γ with the negated conclusion, ¬A. I’ll write ‘∣A∣’ to denote the set of all scenarios
which represent that A is true, and ‘∣Γ∣’ for the set ⋂ A∈Γ ∣A∣, i.e. the set of all
scenarios which represent each member of Γ as being true.6 On the first view, the
content of a deduction Γ ⊢ A comes out as ∣Γ∣ − ∣A∣ (or equivalently, ∣Γ∣ ∩ (∣A∣)c ):
the set of all scenarios which verify each member of Γ, but do not verify A.7
Call this notion content1 . On the second view, the content of Γ ⊢ A comes out
as ∣Γ∣ ∩ ∣¬A∣: the set of all scenarios which verify each member of Γ and also the
negation of A. Call this notion content2 . When negation behaves classically at
all scenarios (i.e. when each s verifies ¬A just in case it does not verify A) these
two notions of content are equivalent: the set of logically possible scenarios which
verify Γ but not A is precisely the set of those which verify both Γ and ¬A.
Our problem is then that, on either notion of content, there are no such
(classical) logically possible scenarios whenever A can be derived logically from Γ.
Hence each notion of content is empty, and the valid deduction Γ ⊢ A is treated
as utterly uninformative. To resolve the puzzle, we require, at a minimum, a
notion of a scenario which allows scenarios which verify Γ and either verify ¬A
or fail to verify A, even when Γ entails A. I’ll assume throughout that we are
interested in a classical notion of entailment (so that ‘valid deduction’ will mean a
classically valid one). If we want to give a non-empty content to such deductions,
therefore, we will need to appeal to non-classical scenarios.8 This task is relatively
4. There are many other notions of content available. Throughout, I will take ‘content’ to mean the
specific epistemic notion I’ve just described.
5. I won’t make anything of the distinction between a specific procedure of deducing A from Γ and
the statement that A can be deduced from Γ, ‘Γ ⊢ A’. I’ll talk throughout in terms of the content of
a deduction Γ ⊢ A. In saying that Γ ⊢ A is informative, I mean that performing some derivation of
A from Γ has the potential to be informative to some agent (or: it would be informative to an agent
who begins with no declarative information at all). Thus, Γ ⊢ A is informative only if its content is
non-empty.
6. Throughout, A, B, C etc. are names for sentences. So I talk about a scenario representing that A is
true, rather than representing that A. I’ll also say that a scenario s verifies A, merely to abbreviate
‘s represents that A is true’. I’ll use the convention that all logical formulae come with implicit
Quine-quotes. Thus ¬A abbreviates ⌜¬A⌝.
7. Here, (∣A∣)c is the set-theoretic compliment of ∣A∣ on the set of all scenarios.
8. Once we admit non-classical scenarios, content1 and content2 may come apart, for there may be
scenarios which verify neither A nor ¬A or both A and ¬A (although see §2 for an argument that
2
easy, for many such structures are well-understood in contemporary model theory.
The harder task is to explain why such structures should count as (or are good
representations of) epistemically or doxastically possible scenarios. To provide a
suitably epistemic notion of content, in terms of which we can say what it is for a
deduction to be informative, the scenarios we use must all represent epistemic (or
doxastic) possibilities: they must represent what seem to be genuine possibilities.9
The remainder of the paper proceeds as follows. In §§2–3 I present and discuss
two accounts of non-classical scenarios which, it has been suggested, are suitable
for our purposes. I argue that neither notion is suitable for our purposes. In §4,
I discuss an analogy between our problem and the sorites paradox, concluding
that the notion of content we are seeking is inherently vague (and as such, we
should reject attempts to model contents by imposing artificial precision). Then,
in §5, I present a formal model of content which treats some (but not all) valid
deductions as being informative. I evaluate the model in §6.
2 The Relational Models Approach
To resolve our problem, we require scenarios which may verify both Γ and ¬A, or
alternatively verify Γ but not verify A, even when Γ ⊢ A. A popular place to look
for such a notion is the model theory of paraconsistent and paracomplete logics,
in which negation behaves in non-classical ways.10 In a model of a paraconsistent
logic, a sentence may be both true and false; in a model of a paracomplete logic, a
sentence may be neither true nor false. In this section, I’ll evaluate the suggestion
that such models provide a way to model the content of deduction.
I’ll call such models relational models, because we obtain them by replacing
the classical valuation function with a relation V between models, sentences and
the classical truth-values, so that a model may relate a sentence to either, both or
neither of the values 1 and 0 (standing for truth and falsity).11 Truth-at-a-model
(⊧) is then defined as follows:
there are no genuinely epistemic scenarios of the latter kind).
9. Note that allowing logically impossible scenarios to play an epistemic role does not entail that
one can know impossibilities (e.g., true contradictions). What one knows is a matter of what is the
case in all accessible epistemic scenarios. Standardly in epistemic logic, we count the actual world as
an epistemic scenario for all agents. Then, if there are no actual true contradictions, no one will be
modelled as knowing any true contradiction.
10. Priest (1987; 2008) gives the background to these logics. Models of these logics have a history
in the epistemic logic literature, particularly in connection with worries about logical omniscience
(Levesque 1984; Lakemeyer 1986; 1987; 1990).
11. There are alternative model theories for paraconsistent and paracomplete logics, which I won’t go
into here.
3
s⊧p
iff
V s p1
s ⊧ ¬p
s ⊧ ¬¬A
iff
iff
V s p0
s⊧A
s ⊧ A∧B
s ⊧ ¬(A ∧ B)
iff
iff
s⊧A&s⊧B
s ⊧ ¬A or s ⊧ ¬B
s ⊧ A∨B
iff
s ⊧ A or s ⊧ B
s ⊧ ¬(A ∨ B)
s⊧A→B
iff
iff
s ⊧ ¬A & s ⊧ ¬B
s ⊧ ¬A or s ⊧ B
s ⊧ ¬(A → B)
iff
s ⊧ A & s ⊧ ¬B
We use relational models to give an account of the content of a classically
valid deduction by allowing relational models to count as scenarios. To see how
this helps, let’s focus on a classical deduction involving some modus ponens steps,
from A → B and A to B. We can find relational models where both premises
are true but B is not. Suppose A is both true and false, and B is (just) false, at s.
Then we have s ⊧ A and s ⊧ ¬A and hence s ⊧ A → B, but not s ⊧ B. Thus the
content1 of the inference from A → B and A to B, defined as (∣A → B∣ ∩ ∣A∣) − ∣B∣,
is non-empty. We also have s ⊧ ¬B, and so the content2 of the inference, defined as
∣A → B∣ ∩ ∣A∣ ∩ ∣¬B∣, is also non-empty. In this way, by allowing relational models
to count as scenarios, we can provide a non-empty content1 and content2 for
some classically valid deductions.
On this picture, not all valid deductions come out as being contentful1 . The
deduction A, B ⊢ A ∧ B remains contentless1 , since any scenario verifying both
A and B also verifies A ∧ B (and so ∣{A, B}∣ − ∣A ∧ B∣ is empty). Indeed, any
classically valid inference not involving ‘¬’ or ‘→’ will be deemed contentless1 , on
this view. This is a bad consequence for an account of content. Modus ponens and
conjunction elimination, for example, do not seem to be wholly different kinds of
inference rule. If one is deemed trivial or obvious, then the other should be too.
One kind of inference should not be deemed to have content if the other does not.
Does our notion of content2 , defined as ∣Γ∣ ∩ ∣¬A∣ for premises Γ and conclusion
A, fare any better? It is easy to see that, when the set of scenarios includes relational
models, every valid inference is deemed contentful2 . Take our conjunction
introduction case from above (which was not deemed contentful1 ). If scenario s
verifies both A and B and in addition verifies either ¬A or ¬B, then it will also verify
both A ∧ B and ¬(A ∧ B). Such scenarios comprise the set ∣A∣ ∩ ∣B∣ ∩ ∣¬(A ∧ B)∣,
and hence constitute the content2 of A, B ⊢ A ∧ B. Even the most trivial inference
of all, from A to A, is deemed contentful2 : ∣A∣ ∩ ∣¬A∣ is the set of all ‘glutty’
scenarios which verify both A and ¬A. Since any sentence A whatsoever can be
represented as being both true and false by a scenario, ∣Γ∣ ∩ ∣¬A∣ is guaranteed to
be non-empty, regardless of any deductive relationships between Γ and A. This
too is a bad consequence for a notion of content. I do not know of any reason for
taking A ⊢ A (for example) to be informative. So, on either notion of content, the
relational models approach does not provide a good account of the content of a
4
valid deduction.
There is a deeper problem with the relational models approach: it fails to
explain why the scenarios it provides are suitable tools for analysing epistemic
notions of content and information. It is a consequence of the account that both
the content1 and content2 of a valid deduction Γ ⊢ A can contain only glutty
models, which assign both 0 and 1 to some sentence.12 Yet it is a constraint on
epistemic scenario-hood that what a scenario represents as being the case must at
least seem possible. We want to model an agent’s learning that A in terms of her
ruling out some scenarios. If what those scenarios represent as being the case is
obviously impossible to any agent who meets minimal standards of rationality,
then there’s no sense in which ruling out those scenarios corresponds to gaining
new information.
From a classical point-of-view, it is obviously, trivially impossible for both
A and ¬A to be true. Dialethists will disagree, of course; and such debates
can’t be settled here. We can agree to fix our attention on agents who take it
to be obviously, trivially impossible for both A and ¬A to be true.13 It might
even be partially constitutive of their notion of being a rational agent that one
takes it to be obviously, trivially impossible for both A and ¬A to be true. We
can build this principle into our notion of informativeness for those agents,
so that any representation of both A and ¬A is ruled out from the start.14
Consequentially, in modelling what’s informative for those agents, we shouldn’t
treat such representations as scenarios at all.15 Once we slim down our class of
(classical plus relational) scenarios in this way, we throw out all the relational
models and are left only with classical models. As we have already seen, none of
these scenarios allows us to provide a non-empty content1 or content2 for any
classically valid inference. Expanding our stock of scenarios with relational models
has not helped us to provide non-empty contents for classically valid inferences.
This is the very feature which makes our problem difficult. If we are to model
the content of a valid deduction as a set of scenarios, then we have to admit
‘impossible’ (non-classical) scenarios. But trivially impossible models of explicit
contradictions cannot feature in any account of rational (but non-ideal) attitudes.
And on the relational models account of deductive content, the trivially impossible
12. To see why, assume that Γ ⊢ A. Then on any consistent truth-assignment to the non-logical
vocabulary of Γ ∪ {A} on which all members of Γ are assigned 1, A will be assigned 1 as well. (This
holds regardless of whether the truth-assignment as a whole is classical or paracomplete, i.e., allowing
that some sentences are assigned neither 0 nor 1.) No scenario in ∣Γ∣ − ∣A∣ or ∣Γ∣ ∩ ∣¬A∣ corresponds
to such a truth-assignment. Hence all scenarios in ∣Γ∣ − ∣A∣ and ∣Γ∣ ∩ ∣¬A∣ (content1 and content2 ,
respectively) assign both 0 and 1 to some sentence in the non-logical vocabulary of Γ ∪ {A}.
13. Indeed, we could even work with agents who allow that it’s possible for both A and ¬A to be true,
but for whom it is obviously, trivially not in fact the case that both A and ¬A are true.
14. Of course, obviousness is both highly subjective and a matter of degree. This does not affect
the argument, since an explicit contradiction A, ¬A represents a determinate case of an obvious
impossibility for all for the agents under consideration.
15. We must be precise when stating this objection. It is not that the relational models approach is
incompatible with a rational agent’s belief in each instance of ¬(A ∧ ¬A). For this is paraconsistently
equivalent to A ∨ ¬A, which is verified by any scenario which assigns some value to A. The objection
is not that the approach conflicts with what rational agents do in fact believe. Rather, the objection
is that, given the uses to which scenarios will be put, the very notion of a scenario excludes such
obviously glutty models.
5
models are all we’re left with. In short, our question is difficult because it requires
us to find scenarios which are impossible, but not trivially so.
In this section, I’ve argued that, when we interpret scenarios as relational
models, neither content1 nor content2 provides a satisfying account of how a valid
deduction can be informative. I then argued that (at least with respect to a certain
background notion of rationality) we shouldn’t count any model of an explicit
contradiction as an epistemic scenario. If so, then both the content1 and content2
of any valid deduction comes out empty, on the relational models approach. In the
next section, I consider an attempt to make sense of models which are impossible,
but subtly so.
3 The Urn Model Approach
In this section, I discuss an attempt by Hintikka (1975) and Rantala (1975)
to provide formal models which ‘look possible but which contain hidden
contradictions’ (Hintikka 1975, 476). They aim to characterise models ‘so subtly
inconsistent that the inconsistency could not be known (perceived) by an everyday
logician, however competent’ (Hintikka 1975, 478). As we saw in the previous
section, relational models are not appropriate in an account of epistemic content,
precisely because whenever they represent the impossible, they represent the
trivially impossible. If Hintikka and Rantala are successful in their aim, then we
have good candidates to play the role of scenarios in our account of content.
The models Hintikka and Rantala describe are based on Hintikka’s (1973a;
1973b) game-theoretic models of classical first-order logic.16 The world is viewed
as an urn from which individuals are drawn by two players, called ‘∀’ and ‘∃’.
In a game G(∀xA), player ∀ must pick an individual from the urn satisfying
A; if she picks individual a, the game continues as G(A[a/x]). Similarly, the
game G(∃xA) requires ∃ to pick an individual a satisfying A and continues as
G(A[a/x]).17 In this way, nested quantifiers represent constraints on sequences
of draws from the urn. Just as in probability theory, individuals can but need
not be replaced after being drawn from the urn. Models in which individuals are
always replaced immediately after being drawn are the invariant models; all others
are changing models. Invariant models provide a classical first-order semantics,
whereas changing models are non-classical (and correspond to what Hintikka
(1975) terms the exclusive interpretation of the quantifiers).
Hintikka’s idea is that, given a sentence of certain game-theoretic complexity,
there is a set of changing models ‘which vary so subtly as to be indistinguishable
from invariant [i.e., classical] ones at a certain level of logical analysis’ (Hintikka
1975, 483). Hintikka (somewhat unfortunately) calls such models ‘impossible
possible worlds’. For the formal details, Hintikka draws on Rantala’s (1975)
notion of an urn model:
16. Hintikka’s models develop the ideas of Henkin (1961) and Peirce (1992).
17. Analogously, ∀ decides whether G(A ∧ B) proceeds as G(A) or as G(B) whereas ∃ decides how
G(A ∨ B) should proceed. The game G(¬A) proceeds as the inverse game G(A), in which the players
swap roles.
6
Definition 1 (Urn sequence) An urn sequence ∆ over a domain D is a countable
sequence ⟨Di ∣ i ∈ N⟩ where D1 = D and, for i ≥ 1, Di ⊆ D i (the ith Cartesian
power of the domain D) such that, for some a′ ∈ D:
⟨a1 ⋯ a i ⟩ ∈ Di only if ⟨a1 ⋯ a i a′ ⟩ ∈ Di+1
and, for all a′ ∈ D:
⟨a1 ⋯ a i ⟩ ∈ Di if ⟨a1 ⋯ a i a′ ⟩ ∈ Di+1 .
Definition 2 (Urn model) An urn model M is a pair ⟨M, ∆⟩, where M is a classical
first-order model with domain D and ∆ is an urn sequence over D. An urn model
M satisfies a sentence A, M ⊧u A, when ∃ has a winning strategy in the game
G(A) played in M.
Definition 3 (d-invariant models) Let δi = {a i ∣ ∃a1 ⋯ ∃a i−1 ⟨a1 ⋯ a i−1 a i ⟩ ∈ Di }.
For any d ∈ N, an urn-model M = ⟨M, ⟨D1 D2 ⋯ ⟩⟩ is d-invariant iff D1 = δ1 =
δ2 = ⋯ = δd .
Here, δi is the set of individuals available at draw i. The d-invariant models behave
as classical models for all sentences A with quantifier depth no greater than d.18
For such A, if M = ⟨M, ∆⟩ is d-invariant, then M ⊧u A iff M classically satisfies A.
Yet a valid sentence with quantifier depth d need not be satisfied by a d ′ -invariant
model, for any d ′ < d.
Unlike relational models, changing urn models do not verify classically
unsatisfiable sentences only at the cost of making some sentences both true and
false (the non-subtle way!). The approach also comes with a well-defined, nontrivial notion of consequence and an accompanying proof-theory (Hintikka 1970;
1973b).19 That’s good news if one wants to investigate the logic of epistemic
notions.20 So urn models seem to be both logically respectable and good candidates
for playing the role of epistemic scenarios.
There are, however, several serious problems with using urn models in this
way. The first problem is this. In using these models, Hintikka works with a
notion of an agent’s logical competence, measured in terms of quantifier depth d.
The rough idea is that greater competence correlates with the ability to reason
correctly with sentences involving higher numbers of embedded quantifiers.21
We could, if we want, replace Hintikka’s individualistic notion of competency
with a communal one, reflecting standards of linguistic understanding, or our
18. The quantifier depth of A is the greatest n such that there is a quantifier in A embedded within
n − 1 other quantifiers (or 0 if A is quantifier-free).
19. Sequoiah-Grayson (2008) discusses Hintikka’s proof theory in detail, and comes to conclusions
similar to those offered below.
20. As I’ll point out in §4, however, the assumption that epistemic contents are closed under logical
rules (other than identity) is rather dubious.
21. In an epistemic logic setting, the epistemic accessibility relation of an agent with competency d is
constrained so that all accessible urn models are d ′ -invariant, for all d ′ ≤ d (Hintikka 1975). That
agent is then modelled as an ideal agent up until the limits of her competency, but no further.
7
communal expectations on rational (but non-ideal) agency.22 But quantifier depth
does not seem to be a very good measure of either logical competence or communal
standards of (non-ideal) rationality. Suppose Anna completes a (correct) proof,
in which no sentence has a quantifier depth greater than d, of a mathematical
statement A. If she completed the proof through skill and not random luck, her
achievement reflects her competence, and so we must assign her a competence
of at least d. Then, no d ′ -invariant model was ever an epistemic possibility for
Anna, for any d ′ ≤ d. But since A appears in the proof, it must have a quantifier
depth no greater than d, and so is true in all d-invariant models. Consequently, in
a Hintikka-style model of knowledge, Anna is modelled as having known A all
along; and in a model of content, her proof is modelled as being contentless and
uninformative. This is just what we want to avoid.
This objection shows only that we shouldn’t link the d parameter directly
to an agent’s competence (or communal standards of competence). Let’s grant,
for the sake of argument, that d can be fixed meaningfully in some other way,
so as to avoid the objection. Even then, a further problem remains. Given that
δ1 = D in any urn model with domain D, we can show that, for any A which
contains no embedded quantifiers and any urn model M = ⟨M, ∆⟩, M ⊧u A iff M
classically satisfies A (Rantala 1975, 466, theorem 1). As a consequence, agents
are modelled as being logically omniscient with respect to all such sentences; and
all inferences involving only such sentences are deemed contentless. This gets
things wrong: if valid deductions can be informative at all, then surely at least
some purely truth-functional (quantifier-free) inferences are informative. It’s not as
if the first-year logic class struggles only with those inferences involving embedded
quantification!
Hintikka’s use of urn models represents an improvement (in certain respects)
on using relational models to capture epistemic scenarios, but it renders too
many inferences uninformative. The root of problem is that using urn models
in this way establishes an absolute cut-off point between potentially informative
and necessarily uninformative inferences (depending on whether they involve
embedded quantifiers). But (as I argue in the next section) in reality there is
no such absolute cut-off point. This is because content, and hence whether an
inference is informative or not, is an inherently vague matter.
4 A Diagnosis of the Problem
In this section, I provide a diagnosis of the problem of how a valid deduction can
be informative. This will pave the way for a formal model, in §5, of the content
of deduction.
There is something very counterintuitive in the claim that the deductive move
from, say, A ∧ B to A is informative. It is tempting to say that anyone who claims to
believe that A ∧ B but not that A is in some way irrational (or confused about what
22. The idea, very roughly, would be that persistent logical mistakes falling below the standard would
reflect the failure to grasp the relevant concept fully, whereas mistakes persistently over the threshold
would instead reflect errors in calculation, lack of cognitive resources and the like.
8
those words mean). But if we assume (perhaps as some principle of rationality)
that inferences made using such rules are wholly uninformative, we will quickly
run into problems, for a deductive proof is no more than the repeated application
of such rules. So if we grant the assumption, then we are at risk of incorrectly
treating any proof whatsoever as being uninformative and hence incapable of
adding to one’s knowledge.23 That is our puzzle.
The case is analogous to a sorites series of colour samples, going gradually
from red to yellow. Adjacent samples are indistinguishable in colour and so it
seems that, if we judge any sample to be red, we should also judge the next one to
be red, too. But, as the first is clearly red, we are then at risk of judging them all
to be red, which is clearly wrong. One could always stipulate that ‘red’ applies
only to the first 23 samples (say), and ‘non-red’ to all the others. But that does
nothing to resolve the puzzle, which concerns our concept of redness and not some
artificial precisification of it. The puzzle is to make sense of truth and inference
in a vague language, so that not every colour sample is counted as being red.
The task is not to reform the language by removing vague predicates. I want to
draw the same moral in our case of deduction: the task is to make sense of a
notion of content such that some, but not all, valid inferences are informative,
without drawing an artificially sharp line between those that are and those that
are not. The normative notion of content we want is a vague notion, precisely
because chains of seemingly uninformative inferences can give rise to informative
deductions.
The analogy with sorites cases highlights another important feature of content.
There must be some relation between content and meaning. In the case of a
conjunction A ∧ B, there must be some relation between the content of A ∧ B, the
contents of A and B, and the meaning of ‘∧’. Since there is a clear link between
the meaning of the logical constants and the proof rules which govern their use,
there must be some link between the content of A ∧ B and the rules we use to
make inferences to and from A ∧ B.24 There is something undeniable about this:
any account which wilfully ignores these rules can’t claim to be a genuine account
of epistemic content for rational agents.
The problematic assumption is that the way to capture such rules is in terms
of closure conditions on truth-at-a-scenario. This forces us into the classical (or
relational) picture of content, on which A ∧ B is true at a scenario s iff both A
and B are true at s. But the classical picture cannot accept that a valid inference
may be informative (§1), whereas the relational models account is not acceptable
for other reasons (§2). To give a non-empty content to valid deductions, we must
avoid those closure conditions; yet we do not want to lose sight completely of
meaning-governing inference rules. Our account of content must be subject to
23. Dummett makes a similar point: ‘When we contemplate the simplest basic forms of inference,
the gap between recognising the truth of the premisses and recognising that of the conclusion seems
infinitesimal; but, when we contemplate the wealth and complexity of number-theoretic theorems
which, by chains of such inferences, can be proved . . . we are struck by the difficulty of establishing
them and the surprises they yield.’ (1978, 297)
24. Note that I’m not claiming that proof rules are constitutive of a logical constant’s meaning. I want
to remain neutral on this issue. It may be that a logical constant’s meaning is fixed by truth-conditions,
which also fix the correct proof rules, for example.
9
those inference rules, even if truth-at-a-scenario is not closed under them. How
can these requirements be satisfied simultaneously?
In the next section, I develop an account of the content of deduction on which
contents are vague, rather than precise. The account avoids the closure conditions
on truth-at-a-scenario which would lead to the classical (or relational model)
account of content. Yet, I will claim, it gives us a notion of content that respects
the inference rules which correspond to the meanings of the logical connectives.
5 A Model of Content
In this section, I set out a formal model of epistemic content (including the content
of deduction). I’ll begin by talking in terms of points, rather than scenarios. Some
but not all of these points will count as epistemic scenarios, and content will be
built from those scenarios. The guiding idea is that, in certain cases, it will be
indeterminate whether a given point counts as a scenario. This will allow us to
define contents which are themselves indeterminate.
I’ll assume a space W of very fine-grained points. For each pair of arbitrary
sets of sentences Γ, ∆, there is a point w ∈ W such that the truths and falsities
according to w are precisely the members of ∆ and Γ, respectively. We define a
model of content M (relative to a language L) to be a tuple
⟨W , R, V − , V + ⟩
where W is a set of fine-grained points, R ⊆ W × 2W is an irreflexive, asymmetric
and non-transitive relation and V − , V + ∶ W Ð→ 2L . I’ll say that a point w verifies
A iff A ∈ Vw+ and falsifies A iff A ∈ Vw− .
The relation R, which relates points to sets of points, captures the structure of
proofs, relative to some fixed set of proof rules. For simplicity, I’ll work with a
classical sequent-style system, but this isn’t essential.25 I’ll focus on the simple case
of a propositional language; extension to a first-order language is easy. Standard
presentations of the sequent calculus focus on sequents, of the form Γ ⊢ ∆, where Γ
and ∆ are multisets of sentences.26 Rules manipulate such sequents, and may have
either one or two sequents as premises (or upper sequents) and a single sequent as
conclusions (or lower sequents). A standard sequent calculus for classical logic
employs left and right logical rules for each connective (see, e.g., Buss 1998), plus
the contraction, identity and cut rules:
Γ, A, A ⊢ ∆
Γ, A ⊢ ∆
Γ, A ⊢ A, ∆
[id]
Γ ⊢ A, A, ∆
Γ ⊢ A, ∆
[cl]
[cr]
Γ ⊢ ∆, A
A, Γ ⊢ ∆
Γ⊢∆
[cut]
25. We can also encode natural deduction systems (Jago 2009c). There is no restriction to classical
systems, either. We can even encode non-monotonic inference rules (Jago 2009a).
26. In a multiset, each element may occur one or more times. We can treat a multiset Γ as a standard
set ∆ coupled with a function # ∶ ∆ Ð→ N, with #x telling us how many occurrences of x appear in Γ.
10
A proof of a sequent Γ ⊢ ∆ is a tree of sequents whose root is Γ ⊢ ∆ and whose leaves
are all instances of id. Any sequent provable using cut can be proved without
it and so I’ll set cut to one side. In practice, proofs are constructed bottom-up,
beginning with the sequent to be proved and working upwards, applying the rules
from lower sequent to upper sequent(s).
I’ll work with a slightly non-standard system, in which (i) Γ and ∆ in a sequent
Γ ⊢ ∆ are sets, rather than multisets of sentences, and (ii) all sentences appearing
in the lower sequent of a rule must appear in the upper sequent(s) too. The logical
rules for ‘¬’ and ‘∨’, for example, are:
Γ, ¬A ⊢ A, ∆
Γ, ¬A ⊢ ∆
Γ, A ⊢ ¬A, ∆
Γ ⊢ ¬A, ∆
[¬l]
Γ, A ∨ B, A ⊢ ∆
Γ, A ∨ B, B ⊢ ∆
Γ, A ∨ B ⊢ ∆
[∨l]
[¬r]
Γ ⊢ A, B, A ∨ B, ∆
Γ ⊢ A ∨ B, ∆
[∨r]
The rules for the other connectives are similarly modified from the standard
logical rules. In practise, these rules require us to carry all sentences with us as
we construct our tree from root upwards. In this system, we can dispense with
contraction (the cl and cr) rules. The system I will work with consists of just these
left and right logical rules for each connective plus id, with no structural rules.
Our proof rules are thus all strong inference rules, in the sense of Buss (1998).
This system is equivalent to the standard presentation: any sequent derivable in
one system is derivable in the other. Hence the new system is sound and complete
with respect to the truth-table semantics.
For every sequent Γ ⊢ ∆, there is a corresponding point w which verifies each
sentence of Γ and falsifies each sentence of ∆. We can, therefore, write that sequent
as Vw+ ⊢ Vw− . This sequent is valid iff w is logically incoherent, for Γ ⊢ ∆ iff the
combined truth of each A ∈ Γ is incompatible with the combined falsity of each
B ∈ ∆. R is then the set of all pairs (w, {u}) for which there are points w and u
and a rule instance r1 :
Vu+ ⊢ Vu−
Vw+ ⊢ Vw−
plus all pairs (w, {u, v}) for which there are points w, u and v and a rule instance
r2 :
Vu+ ⊢ Vu−
Vv+ ⊢ Vv−
+
Vw ⊢ Vw−
In this way, R relates points corresponding to a rule instance’s lower sequent to the
set of points corresponding to its upper sequent(s). Each R-transition corresponds
to a potential inference, all of which involve logical (as opposed to structural)
rules. R as a whole captures all the potential proofs available to us.
Let a point-graph G be any rooted, directed acyclic graph with vertices VG ⊆ W
and edges EG ⊆ W 2 , restricted so that (w, u) ∈ EG only if:
∃X ⊆ W ((w, X) ∈ R & u ∈ X & ∀x(x ∈ X ↔ (w, x) ∈ EG )).
11
In a point-graph G, each vertex has zero, one or two children. Vertices with zero
children are leaves. If w’s only child is u, then there is an instance of a proof
rule such as r1 above. If w has two children u, and v, then there is an instance
of a proof rule such as r2 above. A point w is closed iff Vw+ ∩ Vw− ≠ ∅, and a
point-proof is a point-graph all of whose leaf nodes are closed. The size of a
point-graph G is the number of non-leaf vertices it contains (which corresponds
to the number of inference tokens G captures). Each closed point w is an instance
of the id rule, and each point-proof has an instance of id at each of its leaves.
Given the soundness of our rules and the way point-proofs are constructed from
rule instances, it follows that for any point-proof P, the point w at the root of
P is logically inconsistent (and so the sequent Vw+ ⊢ Vw− is valid). Conversely, for
any point w, if w is logically inconsistent then, given the completeness of the
proof rules and the way point-proofs are constructed by rule instances, there is a
point-proof P with w at its root.
Next, we totally order all our points. Let f w be the size of the smallest pointproof with w at its root, if there is such a point-proof (and undefined otherwise).
Now set w ⪯ w ′ iff either f w ′ is undefined or f w ≤ f w ′ . Points w such that
Vw− ∩ Vw+ ≠ ∅ are ⪯-minimal elements, associated with point-proofs of size 0; I’ll
call these the -points. Points not at the root of any point-proof are ⪯-maximal.
The maximal points are consistent (with respect to our chosen proof rules). All
other points are inconsistent; but not all inconsistent points are on a par. Intuitively,
⪯ orders points by how easy (in terms of proof length) it is to refute that point.
When w ⪯ w ′ , the inconsistencies in w ′ are buried at least as deep (in terms of
proof length) as those in w. So, if one accepts w as an epistemic scenario, then
one should also accept w ′ as an epistemic scenario.
This does not yet give us an account of which points count as epistemic
scenarios, which we need for our account of content. The ordering ⪯ is supposed
to correspond to the way we can order colour samples by their degree of redness.
The problem in that case is that we can’t detect just which set of samples count
as the extension of ‘red’. The colour case and the deduction case are structurally
similar, and so I want to treat them in the same way. Whatever one thinks is the
correct philosophical account of vagueness can be ‘plugged in’ at this point.
I’m tempted by the view that vague cases are cases for which there is no fact
of the matter either way. On this view, we partition the points into three classes:
those that are scenarios, those that are not, and those for which there is no fact of
the matter either way. This partition is constrained by ⪯: if w is a scenario, then
so are all w ′ such that w ⪯ w ′ ; if w is not a scenario, then so are all w ′ such that
w ′ ⪯ w; and if there are no facts of the matter regarding w1 and w2 , and w1 ⪯ w2 ,
then there is no fact of the matter regarding any w ′ such that w1 ⪯ w ′ ⪯ w2 .
One will reason about this set-up using a 3-valued logic, e.g. strong Kleene logic.
Epistemicists, by contrast, will partition points into two: the scenarios and the
non-scenarios, again constrained by ⪯ in the obvious way. Many-valued accounts
will assign a degree of truth δw ∈ [0, 1] to ‘w is a scenario’, constrained so that
δw ≤ δw′ iff w ⪯ w ′ .27
27. For an account along these lines, see Jago 2009b.
12
Epistemic contents inherit the vagueness of ‘epistemic scenario’. Let ∣A∣+ be the
set of all epistemic scenarios which verify A, and ∣A∣− be the set of all epistemic
scenarios which falsify A.28 These sets have indeterminate membership: s ∈ ∣A∣+
iff (i) s verifies A and (ii) s is a scenario. Since it may be indeterminate whether
s is a scenario, it may also be indeterminate whether s ∈ ∣A∣+ (and similarly for
∣A∣− ). Just how this vagueness in content is modelled depends, once again, on
one’s semantics for vagueness.29 The content of A is then the pair (∣A∣+ , ∣A∣− ).
Content thus tracks all scenarios which falsify, as well as all those which verify,
the sentence in question. Finally, I define the content of the deduction from Γ to
A as the set of all epistemic scenarios which verify each of Γ and falsify A, i.e.
∣Γ∣+ ∩ ∣A∣− .30
In this section, I’ve presented models which can be used to define epistemic
contents of sentences and of deductions. Those contents are vague: it may be
indeterminate whether a content has a given point w as a member. Just how this
vagueness is modelled depends on one’s semantics for vagueness in general: the
models given here can be supplemented by any of the formal semantics for vague
languages. In the next section, I discuss the features and advantages of this account
of content.
6 Evaluating the Model
In this section, I review some of the advantageous features of the account I’ve just
presented. To begin with, it improves on the urn-models account (§3) in that it
allows purely truth-functional deductions to count as informative. On the view
I’ve presented, whether a given deduction is informative isn’t a matter of whether
it is contained within this or that fragment of the language; rather, it is a matter
of how difficult that inference is.31 It also improves on the relational models
approach (§2) in that no obviously impossible point is counted as an epistemic
scenario, and so no such point features in the content of any deduction.
On the account I’ve given, truly trivial deductions come out contentless and
hence uninformative, just as we want. For example, suppose w verifies each of
28. Note that, in general, ∣A∣− ≠ ∣¬A∣+ .
29. For epistemicists, ∣A∣+ and ∣A∣− are classical sets whose exact membership (when so described) is
unknowable. On the many-valued approach, ∣A∣+ and ∣A∣− are fuzzy sets, whose membership function
is δ from above.
30. This is a variant on content2 , defined as ∣Γ∣+ ∩ ∣¬A∣+ . Defining content as I have done makes
sense even in system where negation behaves in non-standard ways (e.g., in paraconsistent logic).
Working with content1 has bad consequences. Take A, B ⊢ A ∧ B: we find its content1 simply by
finding scenarios which verify both A and B but say nothing about A ∧ B. There exist such scenarios
so long as A and B are mutually consistent (for example, the incomplete but consistent point which
verifies A and B but nothing else counts as a scenario). This is a rather badly-behaved notion of
content. For example, p, q ⊢ p ∧ q counts as contentful1 , whereas p, ¬p ⊢ p ∧ ¬p does not (because
any point verifying p and ¬p does not count as a scenario), despite each inference being an instance of
conjunction introduction. Similarly, A ⊢ A is deemed contentless1 (as it should be), whereas A ⊢ A ∧ A
and A ⊢ A ∨ A are not. This strikes me as a bizarre position to hold on content.
31. More precisely: it is a factor of the shortest number of inference steps required to move from
premises to conclusion, relative to some fixed set of inference rules.
13
A → B, A and falsifies B. Then there is a point-proof with w at its root, which we
can represent as
A → B, A ⊢ A, B
A → B, A, B ⊢ B
A → B, A ⊢ B
This graph has size 1, hence f w ≤ 1. No such point counts as an epistemic
scenario, hence the content of the deduction from A → B and A to B, defined as
∣A → B∣+ ∩ ∣A∣+ ∩ ∣B∣− , is empty.
Yet not all valid deductions are empty. Suppose that points w for which f w ≥ m
count as epistemic scenarios. Then the deduction
p1 , p1 → p2 , p2 → p3 , . . . , pn−1 → pn ⊢ pn
is contentful when n > m. Its content consists of scenarios verifying p1 and each
pi → pi+1 (i < n) and falsifying pn . We can verify that there exist such scenarios
as follows. Let w be such that Vw+ = (⋃i<n {pi → pi+1 }) ∪ {p1 } and Vw− = {pn }.
The shortest point-proof with w at its root corresponds to n − 1 applications
of →l, hence f w = n − 1 ≥ m and so, by assumption, w counts as a scenario.
There are then infinitely many scenarios verifying p1 and each pi → pi+1 (i < n)
and falsifying pn : to construct one, simply extend Vw+ or Vw− (or both) in a way
which does not allow for a smaller point-proof to be constructed than the one just
considered.
We have an account of content on which some, but not all, valid inferences
are informative. What of the additional requirement, discussed in §4, that the
content of a logically complex sentence should in some way be linked to the proof
rules governing the use of its main connective? Let’s focus, for the moment, on the
positive component ∣A∣+ of sentence A’s content. In a classical possible-worlds
system of content, we would have that
(1) ∣A ∧ B∣+ ⊆ ∣A∣+ and ∣A → B∣+ ∩ ∣A∣+ ⊆ ∣B∣+ .
These inclusion relationships capture (one aspect of) the meaning of ‘∧’ and ‘→’.
But these inclusion relationships do not hold in our present epistemic system, for
epistemic scenarios are not closed under conjunction elimination or modus ponens.
We can find scenarios w which are members of ∣A ∧ B∣+ but not ∣A∣+ . Nevertheless,
our epistemic notion of content does capture an aspect of these classical inclusion
relationships. The classical possible-worlds framework identifies (∣A∣− )c (the settheoretic complement of ∣A∣− ) with ∣A∣+ and so, on a domain of classical possible
worlds, ∣A∣+ ⊆ ∣B∣+ holds iff ∣A∣+ ⊆ (∣B∣)c holds. Thus on a domain of classical
possible worlds, the inclusion relationships in (1) are equivalent to
(2) ∣A ∧ B∣+ ⊆ (∣A∣− )c and ∣A → B∣+ ∩ ∣A∣+ ⊆ (∣B∣− )c .
14
Although epistemic space does not verify the inclusion relationships in (1), it does
verify those in (2).32 This is one way in which this epistemic notion of content
captures the meanings of ‘∧’ and ‘→’. Similar things can be said for the other
connectives.
In fact, although epistemic space does not validate (1), it nevertheless enforces a
tight relationship between ∣A ∧ B∣+ and ∣A∣+ and between ∣A → B∣+ ∩ ∣A∣+ and ∣B∣+
(and similarly for the other connectives). Let’s focus on the region r of ∣A ∧ B∣+ not
included in ∣A∣+ . Some of the scenarios in that region might be just one inference
away from non-scenarios (or from indeterminate cases of scenarios). Set such
scenarios to one side. For all the remaining scenarios w, there is a further scenario
u ∈ ∣A∣+ such that (w, {u}) ∈ R. Intuitively, such scenarios are as close as they
could be to ∣A∣+ , without actually being in that region. In this sense, ∣A ∧ B∣+
comes as close as it could be to ∣A∣+ without being included in ∣A∣+ . We can say
similar things about the relationships between ∣A → B∣+ ∩ ∣A∣+ and ∣B∣+ , between
∣A∣+ and ∣A ∨ B∣+ , and so on.33 This ‘closeness’ relationship between the relevant
contents justifies the claim that those contents respect the relevant inference rules.
And since those inference rules are intimately connected to the meanings of the
logical constants, this in turn justifies the claim that this model of content respects
the meanings of the logical constants.
In summary, we have a model of content which counts some, but not all, valid
deductions as informative. In particular, completely trivial deductions are modelled
as contentless and hence uninformative. Content (and hence informativeness)
is treated as a vague notion: it is indeterminate just which points constitute a
particular content. Moreover, our model preserves the intimate relation between
the content of A ∧ B and the contents of A and B (and similarly for other
connectives). Those contents are as close as they could be, without collapsing into
the classical picture of content, which is unable to model informative inference.
References
Buss, Samuel. (1998). Handbook of proof theory, volume 137, Elsevier, Amsterdam.
Chalmers, David. (2002). ‘The components of content’, Philosophy of Mind: Classical and
Contemporary Readings, ed. D. Chalmers, Oxford University Press, pp. 608–633.
Chalmers, David. (2010). ‘The nature of epistemic space’, Epistemic Modality, ed. A. Egan
and B. Weatherson, Oxford University Press.
Dummett, Michael. (1978). ‘The justification of deduction’, Truth and other enigmas,
Harvard University Press, Cambridge, MA, pp. 166–185.
32. Of course, it is not the case that ∣A∣+ ⊆ (∣B∣− )c holds whenever ∣A∣+ ⊆ ∣B∣+ holds on a classical
possible-words domain. For some hard-to-prove tautology ⊺, for example, we can have ∣A∣+ ⊈ (∣⊺∣− )c .
33. We cannot say something similar about ∣A∣+ and ∣B∣+ whenever A ⊢ B. For if the shortest proof
from A to B is long, then some scenarios in ∣A∣+ might be quite a long way off (in terms of R-transitions)
from any scenario ∣B∣+ . But this is the point of the epistemic notion of content: as proofs from A to B
become harder to spot, the content of A becomes more remote from the content of B. This is precisely
why such inferences are informative, whereas inferences such as A ∧ B ⊢ A are not.
15
Henkin, Leon. (1961). ‘Some remarks on infinitely long formulas’, Infinitistic Methods,
Pergamon Press, Oxford, pp. 167–183.
Hintikka, Jaakko. (1962). Knowledge and belief: an introduction to the logic of the two
notions, Cornell University Press, Ithaca, N.Y.
Hintikka, Jaakko. (1970). ‘Surface information and depth information’, Information and
Inference, ed. J. Hintikka and P. Suppes, Reidel, Dordrecht.
Hintikka, Jaakko. (1973a). Logic, Language-Games and Information: Kantian Themes in
the Philosophy of Logic, Clarendon Press, Oxford.
Hintikka, Jaakko. (1973b). ‘Surface semantics and its motivation’, Truth, Syntax and
Modality, ed. H. Leblanc, North-Holland, Amsterdam.
Hintikka, Jaakko. (1975). ‘Impossible possible worlds vindicated’, Journal of Philosophical
Logic 4: 475–484.
Jago, Mark. (2009). ‘Epistemic logic for rule-based agents’, Journal of Logic, Language
and Information, 18(1):131–158.
Jago, Mark. (2009). ‘Logical information and epistemic space’, Synthese, 167(2):327–341.
Jago, Mark. (2009). ‘Resources in epistemic logic’, in J.-Y. Béziau and A. Costa-Leite, editors,
Dimensions of Logical Concepts, volume 55, pages 11–33. Coleção CLE, Campinas,
Brazil.
Lakemeyer, Gerhard. (1986). ‘Steps towards a first-order logic of explicit and implicit
belief’, Proceedings of the First Conference on Theoretical Aspects of Reasoning About
Knowledge, ed. J. Y. Halpern, Morgan Kaufmann, San Francisco, pp. 325–340.
Lakemeyer, Gerhard. (1987). ‘Tractable metareasoning in propositional logic of belief’,
Proceedings of the Tenth International Joint Conference on Artificial Intelligence,
pp. 401–408.
Lakemeyer, Gerhard. (1990). ‘A computationally attractive first-order logic of belief’,
Proceedings of JELIA 90, Springer, Heidelberg, pp. 333–347.
Levesque, Hector. (1984). ‘A logic of implicit and explicit belief’, Proceedings of the Fourth
National Conference on Artificial Intelligence, pp. 198–202.
Lewis, David. (1975). ‘Language and languages’, Language, Mind and Knowledge, ed.
K. Gunderson, University of Minnesota Press, pp. 3–35.
Lewis, David. (1986). On the Plurality of Worlds, Blackwell, Oxford.
Peirce, Charles. (1992). Reasoning and the Logic of Things: The Cambridge Conferences
Lectures of 1898, Harvard University Press, Cambridge Mass.
Priest, Graham. (1987). In Contradiction: A Study of the Transconsistent, Martinus Nijhoff,
Dordrecht.
Priest, Graham. (2008). An Introduction to Non-Classical Logic, Cambridge University
Press, Cambridge.
16
Rantala, Veikko. (1975). ‘Urn models’, Journal of Philosophical Logic 4: 455–474.
Sequoiah-Grayson, Sebastian. (2008). ‘The scandal of deduction’, Journal of Philosophical
Logic 37(1): 67–94.
Stalnaker, R. (1976). ‘Propositions’, Issues in the Philosophy of Language, ed. A. MacKay
and D. Merrill, Yale University Press, New Haven, pp. 79–91.
Stalnaker, R. (1984). Inquiry, MIT Press, Cambridge, Mass.
van Benthem, J. (2011). Logical dynamics of information and interaction, Cambridge
University Press, Cambridge.
van Benthem, J. and Martinez, M. (2008). ‘The stories of logic and information’,
Handbook of the Philosophy of Information, ed. J. van Benthem and P. Adriaans,
Elsevier, Amsterdam, pp. 217–280.
17