Mathematical Logic for Students
Mathematical Logic for Students
Adrien Deloro
Spring 2009
Contents
1
1.2.1 Structures . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
1.2.2 Parameters and Interpretations . . . . . . . . . . . . . . . 51
1.2.3 Satisfaction and Semantic Consequence . . . . . . . . . . 51
1.3 Substitutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
1.3.1 Substitutability . . . . . . . . . . . . . . . . . . . . . . . . 53
1.3.2 A Renaming Algorithm . . . . . . . . . . . . . . . . . . . 54
1.3.3 Substitutions and Satisfaction . . . . . . . . . . . . . . . . 55
1.4 Deductions and Soundness . . . . . . . . . . . . . . . . . . . . . 57
1.4.1 Deductions . . . . . . . . . . . . . . . . . . . . . . . . . . 57
1.4.2 Simplifying the Language . . . . . . . . . . . . . . . . . . 58
1.4.3 Soundness . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
1.5 Completeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
1.5.0 Strategy and Witnesses . . . . . . . . . . . . . . . . . . . 63
1.5.1 Expanding the Language . . . . . . . . . . . . . . . . . . 64
1.5.2 Extending the Theory . . . . . . . . . . . . . . . . . . . . 66
1.5.3 Finding a Structure (and Assignment) . . . . . . . . . . . 68
1.6 Consequences and Compactness . . . . . . . . . . . . . . . . . . 70
1.6.1 Decidability* . . . . . . . . . . . . . . . . . . . . . . . . . 70
1.6.2 Compactness . . . . . . . . . . . . . . . . . . . . . . . . . 71
1.6.3 Non-Standard Analysis* . . . . . . . . . . . . . . . . . . . 72
1.7 An Alternate Proof of Compactness* . . . . . . . . . . . . . . . 74
1.7.1 Filters and Ultrafilters* . . . . . . . . . . . . . . . . . . . 74
1.7.2 Ultraproducts (The Łoś Structure)* . . . . . . . . . . . . 76
1.7.3 Alternate Proof of Compactness* . . . . . . . . . . . . . . 78
1.8 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
2
1’.4.2 ∞-isomorphisms . . . . . . . . . . . . . . . . . . . . . . . 106
1’.4.3 Finitary Back-and-Forth* . . . . . . . . . . . . . . . . . . 108
1’.5 Quantifier Elimination and ω-Saturation . . . . . . . . . . . . . . 111
1’.5.1 Elimination Sets . . . . . . . . . . . . . . . . . . . . . . . 111
1’.5.2 ω-Saturated Models . . . . . . . . . . . . . . . . . . . . . 114
1’.5.3 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
1’.6 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
Indices 128
Index of Notions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
Index of Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
3
᾿Εν ἀρχῇ ἦν ὁ λόγος
The Greeks didn’t have a formal language, which in Europe appears during
the Renaissance. Because of our education it now seems very natural to us
that every mathematical statement may be translated into specific signs; yet
this fascinating fact has deeply moved Leibniz’ mind. The adequation of such
a poor alphabet to the manifold hues of mathematics is striking indeed. One
should however not succumb, and their temptations are always present, to the
Sirens of sheer formalism; a mathematician aims at transcribing his intuitions
into symbols, not at assembling symbols and then seeing if they by chance make
any sense. Never does he renounce lucid thinking nor his mother language; and
even less does the logician.
A logician is a mathematician who bears in mind both the mathematical
truths and the way they are stated; able to distinguish the meaning from the
formulation, he can study their interplay. If the mathematician regards language
as a tool to express some interesting properties, the logician thinks the properties
of this tool are themselves worth the interest.
4
Though the objects and language might seem a little less familiar, we shall
follow essentially the same guidelines in our study of various forms of logic:
• agree on a collection of interesting objects
• give a set of basic symbols (an alphabet)
• give a collection of rules for assembling them (a syntax)
• give a meaning to the language (a semantics).
After that, the study of the relationships between the language and what it
stands for can start.
The semester
In the course of the semester we shall study three levels of logic, from the less
to the more expressive.
5
• Propositional Logic concerns itself with compound propositions, using the
five usual connectives. In propositional logic, our basic symbols already
represent propositions; we connect them with the usual connectives. This
form of logic can be dealt with using truth tables: it suffices to specify for
each P is P is true or false. It is not very expressive, but this playground
is an extremely interesting field of investigation for dgging up notions
relevant in deeper logics.
• First-order logic is the natural language for elements. Our basic symbols
are now for elements, and certain (given) relations they can have. One
sees that this is already deeper than propositional logic; one also sees that
one has to introduce quantification on elements. First-order logic is strong
enough to encode much of interesting mathematics, and still enjoys very
nice properties.
• Second-order logic builds on the previous one. Our symbols can now also
be for “predicates‘”, that is for sets, relations. But unlike first-order logic,
second-order logic may quantify on those. So we need new quantifiers
(roughly speaking, we now may quantify on classes, not only on objects).
Second-order logic is extremely expressive, and this is actually the reason
why the beautiful, deep theorems we shall prove in the first two cases will
fail in this somehow too general context.
Looking back at our first steps, we realize that propositional logic is some
kind of zeroeth-order logic. After the investigation of these three forms of logic,
we should believe that the one that best fits our needs is first-order logic. We
shall then move to model theory, which is the mathematical study of first-order
theories. This should explain the rather unorthodox numbering.
6
Chapter 0
Propositional Logic
(Allegro)
In this chapter:
• Describe the language of propositional logic (§0.1)
• Describe the objects of propositional logic (§0.2)
• Formalize the notion of entailment (§0.2.2)
7
0.1 Syntax
We first deal with the familiar syntax of propositional logic. The expressions are
(finite) sequences of well-known symbols (§0.1.1), but not all such expressions
may be given a meaning. This will lead us to the relevant notion of formula
(§0.1.2). A key, and fairly natural, property of the language is that there is
exactly one way to read a relevant formula (Theorem 0.1.9 of §0.1.3).
8
Definition 0.1.6 (theory). A theory is a set of wff’s.
Wff’s are built from the sentence symbols with connectives. As this is a
recursive definition, most of our proofs on formulas will proceed by induction
(notice that the length of a compound proposition is bigger than that of the
propositions which compose it).
Let us be more technical and check that Definition 0.1.4 makes sense.
Notation 0.1.7. We define the following functions.
C¬ : E → E
ϕ →
7 (¬ϕ)
C∧ : E2 → E C∨ : E2 → E
(ϕ, ψ) 7→ (ϕ ∧ ψ) (ϕ, ψ) 7→ (ϕ ∨ ψ)
C→ : E2 → E C↔ : E2 → E
(ϕ, ψ) 7→ (ϕ → ψ) (ϕ, ψ) 7→ (ϕ ↔ ψ)
Notice that these functions actually take WFF (resp. WFF2 ) to WFF.
Notation 0.1.8. Let WFF0 = A, and for n ∈ N,
[ [ [ [
WFFn+1 = WFFn C¬ (WFFn ) C∧ (WFF2n ) · · · C↔ (WFF2n )
It is clear that for all n ∈ N, WFFn is countable. It is also clear that WFF =
∪n∈N WFFn . This proves that WFF is a set, and even a countable set. More
generally, if A is infinite (but perhaps uncountable), then Card WFF = Card A.
Corollary 0.1.10. Let ϕ be a wff of length > 1. Then there is a unique decom-
position of ϕ among the following possibilities:
(This means that the connective and the composing propositions are unique.)
9
Lemma 0.1.11 (balanced parenthesing).
Proof .
(i). We show the property by induction on (the length of) ϕ, using the de-
scription of WFF found above (Notation 0.1.8).
If ϕ ∈ WFF0 , then ϕ is a sentence symbol, and this is obvious. Assume
the result is know for elements of WFFn . Then an element of C¬ (WFFn )
is of the form ϕ = (¬ψ) with ψ ∈ WFFn , so by induction ψ has as many
left parentheses as right parentheses, and so does ϕ. The same holds for
elements of C∧ (WFF2n ), C∨ (WFF2n ), C→ (WFF2n ), and C↔ (WFF2n ). So any
element of WFFn+1 has as many left parentheses as right parentheses.
By induction, the property is true on WFF.
This shows the property for ϕ. One copes with other four connectives in
a similar vein.
(iii). A consequence of (i) and (ii) (the empty expression is not a wff anyway).
Exercise 0.2
(ϕ1 ψ1 ) = (ϕ2 ψ2 )
ϕ1 ψ1 ) = ϕ2 ψ2 )
10
Then ϕ1 or ϕ2 must be an initial segment of the other; without loss of
generality ϕ1 is an initial segment of ϕ2 . If ϕ1 is a proper initial segment of
the wff ϕ2 , then from balanced parenthesing (Lemma 0.1.11 (iii)), ϕ1 can’t be
a wff, a contradiction.
So ϕ1 = ϕ2 , and deleting it we reach
ψ1 ) = ψ2 )
At this point it is clear that = ; deleting it, and the final right parentheses,
we derive
ψ1 = ψ2
Of course the case of C¬ is even easier than that of binary connectives, so
everything is proved.
Such a quick proof of the Unique Readability Theorem is the reason why we
have introduced so many parentheses, in spite of the disgusting redundancies
they create. From now on we freely omit parentheses when there is no ambiguity,
i.e. we write ¬ϕ ∧ ψ instead of ((¬ϕ) ∧ ψ).
End of Lecture 1.
0.2 Semantics
We briefly escape the claws of syntax, and give sense to our symbols. So far
they bore no meaning; at no point yet did we agree on what they stand for.
Once the sentence symbols have been given a truth value, the “natural”
truth value of every wff is known.
11
Proposition 0.2.3. Let v be a truth assignment. Then there is a unique ex-
tension v : WFF → {F, T } such that:
• v(An ) = v(An ) for each n ∈ N;
• v(¬ψ) = T iff v(ψ) = F ;
• v(ψ1 ∧ ψ2 ) = T iff v(ψ1 ) = T and v(ψ2 ) = T ;
• v(ψ1 ∨ ψ2 ) = T iff v(ψ1 ) = T or v(ψ2 ) = T ;
• v(ψ1 → ψ2 ) = T iff v(ψ1 ) = F or v(ψ2 ) = T ;
• v(ψ1 ↔ ψ2 ) = T iff v(ψ1 ) = v(ψ2 ).
Proof . This extension v is well-defined by Unique Readability (Theorem 0.1.9).
A quick induction shows that it is unique.
Example 0.2.4. If v(A1 ) = v(A2 ) = T and v(A3 ) = v(A4 ) = F , then one has
v((A1 → ¬A2 ) ↔ (A3 ∨ A4 )) = T .
Only a little later (at the end of §0.2.3) shall we identify v to is extension v̄.
12
One also says that Σ tautologically implies ϕ. If ∅ |= ϕ, one may then say
that ϕ is a tautology, or that ϕ is valid. We do not insist on classical terminology.
Lemma 0.2.10. Σ ∪ {ϕ} is satisfiable iff Σ 6|= ¬ϕ.
Proof . Suppose that Σ ∪ {ϕ} is satisfiable. Then there is a truth assignment
v satisfying Σ and ϕ; by definition (Proposition 0.2.3), v does not satisfy ¬ϕ.
This means that Σ does not entail ¬ϕ.
Conversely suppose that Σ 6|= ¬ϕ. There is therefore a truth assignement v
satisfying Σ, but not satisfying ¬ϕ. It follows v(ϕ) = T , so v actually satisfies
Σ ∪ {ϕ}.
Theorem 0.2.11. Let Σ0 be a finite theory and ϕ a wff. Then there is an
algorithm which decides whether Σ entails ϕ or not.
Proof . In Σ0 ∪ {ϕ}, only finitely many sentence symbols occur. List all (but
finitely many) truth tables involving them. If all truth tables satisfying Σ0 also
satisfy ϕ, the answer is “yes”; otherwise answer “no” at first flaw.
Question. And what about an infinite set?
This question will find an answer when we can reduce the infinite to the finite
(Compactness, Theorem 0.5.2). The answer is exactly half as good (Corollary
0.5.14). This is far from obvious, as Σ might now involve infinitely many sen-
tence symbols, giving rise to continuum many truth tables. However ϕ is a
single formula, so one has the feeling that some finite subset Σ0 of Σ would
suffice to entail ϕ, in case Σ does.
There is however no bound a priori on how many sentence symbols are
involved in Σ0 , as shows the example ϕ = A1 , Σ = {A2 , A2 , → A3 , . . . , An−1 →
An , An → A1 }. This is actually the reason why the algorithm for infinite Σ
given in Corollary 0.5.14 cannot be implemented to answer “no”: if it hasn’t
answered “yes” after a long waiting time, it could be
• either because Σ |= ϕ, but we haven’t waited long enough to be positive
about it;
• or simply because Σ 6|= ϕ, but we’ll never know.
These questions will be made more precise in §0.5.2.
13
Notation 0.2.12. The collection WFF0 is the smallest set such that:
• each An is a wff;
• if ϕ and ψ are wff, so are ¬ϕ and ϕ → ψ.
The set WFF0 would require a version of the Unique Readability Theorem
(Theorem 0.1.9), which may be regarded as a special case of the previous one,
since WFF0 ⊆ WFF. In any case, we keep working with few parentheses.
There is of course a natural way to translate a “standard” wff ϕ (using five
connectives) into a wff’ ϕ0 using only ¬ and →.
• (ϕ ∧ ψ)0 is ¬(ϕ0 → ¬ψ 0 );
• (ϕ ∨ ψ)0 is (¬ϕ0 → ψ 0 );
• (ϕ → ψ)0 is (ϕ0 → ψ 0 );
• (ϕ ↔ ψ)0 is ¬((ϕ0 → ψ 0 ) → ¬(ψ 0 → ϕ0 )).
Hence for any formula ϕ, the translation ϕ0 uses only two connectives, and
intuitively bears the same meaning.
Example 0.2.14. Let ϕ be (A1 ↔ A2 )∧(A3 ∨A1 ). Then ϕ0 is ¬((A1 ↔ A2 )0 →
¬(A3 ∨ A1 )0 ), which is ¬(¬((A1 → A2 ) → ¬(A2 → A1 )) → ¬(¬A3 → A1 )).
The issue is that this new notion of formula gives rise to another extension
of truth assignment, and therefore to a new notion of entailment. We must
check that these new notions do coincide with their original “five connectives”
analogues.
Notation 0.2.15. For a truth assignment v : WFF → {F, T }, let v 0 be the
extension of v defined by:
14
Proof . By induction on ϕ.
• If ϕ is a sentence symbol, then by definition v(ϕ) = v(ϕ) = v 0 (ϕ0 ).
• Suppose that ϕ is ¬ψ. Then v(ϕ) = T is equivalent to v(ψ) = F , which is
by induction equivalent to v 0 (ψ 0 ) = F , which is equivalent to v 0 (ϕ0 ) = T .
• Suppose that ϕ is ψ1 → ψ2 . Then v(ϕ) = T iff v(ψ1 ) = F or v(ψ2 ) = T
iff v 0 (ψ10 ) = F or v 0 (ψ20 ) = T iff v 0 (ψ10 → ψ20 ) = T iff v 0 (ϕ0 ) = T .
• Suppose that ϕ is ψ1 ∧ ψ2 . Then v(ϕ) = T iff v(ψ1 ) = v(ψ2 ) = T iff
v 0 (ψ10 ) = v 0 (ψ20 ) = T iff v 0 (¬(ψ10 → ¬ψ20 )) = T iff v 0 (ϕ0 ) = T .
• We proceed similarly for the remaining two connectives ∨ and ↔.
Any modification, even apparently harmless, of the notion of truth assign-
ment, induces a modfication of the notion of semantic consequence (Definition
0.2.9).
Notation 0.2.17. Let Θ be a consistent subset of WFF0 and ψ ∈ WFF0 . Write
Θ |=0 ψ if whenever v 0 satisfies Θ, then v 0 (ψ) = T .
Corollary 0.2.18. Σ |= ϕ iff Σ0 |=0 ϕ0 .
Proof . Suppose that Σ |= ϕ. We show that Σ0 |=0 ϕ0 . Let v be a truth
assignement such that v 0 satisfies Σ0 . By Lemma 0.2.16, v satisfies Σ. By
assumption, v satisfies ϕ. Using Lemma 0.2.16 again, v 0 satisfies ϕ0 .
Now suppose that Σ0 |=0 ϕ0 . We show that Σ |= ϕ. Let v be a truth assign-
ment such that v satisfies Σ. By Lemma 0.2.16, v 0 satisfies Σ0 . By assumption,
v 0 satisfies ϕ0 . Using Lemma 0.2.16 again, v satisfies ϕ.
From now on, we know that the semantics induced by two connectives is the
same as that induced by five. In particular we happily forget about v 0 and |=0 .
(We are not entirely done with ϕ0 however, since a similar verification will have
to be carried on the syntactic side, §0.3.3.)
We shall also identify any truth assignment v with its extension v.
End of Lecture 2.
15
there are other logics (for instance, intuitionistic logic, which does not allow
contradiction proofs). They all yield different notions of deduction. We made
the choices which best reflect a normal mathematician’s thought.
After a couple of examples (§0.3.2), we shall proceed in §0.3.3 to reducing
the language to two connectives. There are indeed redundancies as we have
observed (Corollary 0.2.18). We shall show that one may restrict the language
and deduction rules to ¬ and → without affecting the power of the notion of
deduction.
• There is only one axiom (so far). In natural deduction for propositional
logic, the only way to start a proof is by assuming something.
16
negation and contradiction, though it seems satisfactory to us as working
mathematicians.
The elimination rule ¬e is necessary to show the excluded middle law
(Theorem 0.3.11 below); if we introduce a weaker rule, excluded middle
need not hold any more. Exercise 0.10
How hard it may be to believe at first, ¬i is not related to reduction to
absurdum. Only ¬e is.
• There are two elimination rules for ∧ as there are two possible ways to
use a conjunction.
• Dually, there are two introduction rules for ∨, as there are two ways to
prove a disjunction. This is in general difficult, since when trying to show
ϕ ∨ ψ, one never knows a priori which is true; contradiction proofs will be
helpful (see the proof of Theorem 0.3.11 below).
On the other hand ∨e may look subtle at first sight, but it merely describes
the process of a case-division.
We shall start using quantifiers in the next chapters and introduce adequate
rules; for the moment they won’t appear, as we work in Propositional Logic.
Remark 0.3.4. Notice that in our deductions we may have several assumptions,
but at each step we have only one conclusion. This is another typical feature of
Natural Deduction which supposedly reflects the way we mathematicians work.
But there are other, more complex, deduction systems (for instance, Gentzen’s
Sequent Calculus).
A formalization of deduction yields the following essential notion.
Definition 0.3.5 (consistency). A theory Σ is consistent if there is no wff ϕ
such that Σ ` ϕ and Σ ` ¬ϕ.
17
In other words, a theory is consistent if it is not possible to deduce a contra-
diction from it. There is a symbol for inconsistency (or contradiction), which
we shall not use. Exercise 0.9
18
Suppose Σ ` ¬ψ → ¬ϕ. What we just proved implies Σ ` ¬¬ϕ → ¬¬ψ.
Therefore
Ax Ax
{ϕ} ` ϕ {¬ϕ} ` ¬ϕ
Wk Wk
{ϕ, ¬ϕ} ` ϕ {ϕ, ¬ϕ} ` ¬ϕ
¬i
{ϕ} ` ¬¬ϕ Σ ` ¬¬ϕ → ¬¬ψ
Wk WK
Σ ∪ {ϕ} ` ¬¬ϕ Σ ∪ {ϕ} ` ¬¬ϕ → ¬¬ψ
→e
Σ ∪ {ϕ} ` ¬¬ψ
¬e
Σ ∪ {ϕ} ` ψ
→i
Σ`ϕ→ψ
Lemma 0.3.9 (see Lemma 0.2.10). Σ ∪ {ϕ} is not inconsistent iff Σ 6` ¬ϕ.
Proof . Suppose Σ ∪ {ϕ} is not inconsistent and Σ ` ¬ϕ. Then Σ ∪ {ϕ} ` ϕ
(by axiom and weakening), and Σ ∪ {ϕ} ` ¬ϕ (by assumption). It follows that
Σ ∪ {ϕ} is inconsistent, against the assumption. So Σ 6` ¬ϕ.
Suppose that Σ 6` ¬ϕ, and Σ ∪ {ϕ} is inconsistent. By inconsistency , there
is ψ be such that Σ ∪ {ϕ} ` ψ and Σ ∪ {ϕ} ` ¬ψ. Then by ¬i , Σ ` ¬ϕ, against
the assumption. So Σ ∪ {ϕ} is not inconsistent.
Notice that the latter proof did not involve a semantic analog of the elimi-
nation of double negation (¬e ), but only ¬i ; whence the odd statement. As you
believe in ¬e , you may say “is consistent” for “is not inconsistent”.
19
Proof . Due to pagesetting limitations, I must pause and resume.
Ax
{ϕ} ` ϕ
Wk Ax
{¬(ϕ ∨ ¬ϕ), ϕ} ` ϕ {¬(ϕ ∨ ¬ϕ)} ` ¬(ϕ ∨ ¬ϕ)
∨i Wk
{¬(ϕ ∨ ¬ϕ), ϕ} ` ϕ ∨ ¬ϕ {¬(ϕ ∨ ¬ϕ), ϕ} ` ¬(ϕ ∨ ¬ϕ)
¬(ϕ ∨ ¬ϕ) ` ¬ϕ
∨i
¬(ϕ ∨ ¬ϕ) ` ϕ ∨ ¬ϕ
| {z }
(∗)
(∗)
Ax
¬(ϕ ∨ ¬ϕ) ` ϕ ∨ ¬ϕ ¬(ϕ ∨ ¬ϕ) ` ¬(ϕ(∨¬ϕ)
¬i
` ¬¬(ϕ ∨ ¬ϕ)
¬e
` ϕ ∨ ¬ϕ
20
Of course
Σ ` ψ0
induction
Σ`ψ
is a similar shortcut, and is licit for a similar reason.
• Suppose that ϕ is a sentence symbol. Then ϕ0 = ϕ, and the claim is
obvious.
• Suppose that ϕ is ¬ψ. Then ϕ0 = ¬(ψ 0 ). Now
Ax
{ψ 0 } ` ψ 0
Ax Wk
{ϕ} ` ϕ {ϕ, ψ 0 } ` ψ 0
Wk induction
{ϕ, ψ 0 } ` ϕ {ϕ, ψ 0 } ` ψ
¬i
{ϕ} ` ¬ψ 0
|{z}
=ϕ0
and
Ax
{ψ} ` ψ
0 0 Ax Wk
{ϕ } ` ϕ {ϕ0 , ψ} ` ψ
Wk induction
{ϕ0 , ψ} ` ϕ0 {ϕ0 , ψ} ` ψ 0
¬i
{ϕ} ` ¬ψ
|{z}
=ϕ0
Of course the case of ¬ and → was trivial. The real proof starts here.
• Suppose that ϕ = ψ1 ∧ ψ2 , so that ϕ0 = (¬(ψ10 → ¬ψ20 )).
21
– One has
Ax
{ϕ} ` ϕ
Wk
{ϕ, ψ10 → ¬ψ20 } ` ϕ
∧e Ax
{ϕ, ψ10 → ¬ψ20 } ` ψ1 {ψ10 → ¬ψ20 } ` ψ10 → ¬ψ20
induction Wk
{ϕ, ψ10 → ¬ψ20 } ` ψ10 {ϕ, ψ10 → ¬ψ20 } ` ψ10 → ¬ψ20
→e
{ϕ, ψ10 → ¬ψ20 } ` ¬ψ20
| {z }
(∗)
Hence
Ax
{ϕ} ` ϕ
Wk
{ϕ, ψ10 → ¬ψ20 } ` ϕ
∧e
(∗) {ϕ, ψ10 → ¬ψ20 } ` ψ2
induction
{ϕ, ψ10 → ¬ψ20 } ` ¬ψ20 {ϕ, ψ10 → ¬ψ20 } ` ψ20
¬i
{ϕ} ` ¬(ψ10 → ¬ψ20 )
Therefore
Ax
(†) {ϕ0 } ` ¬(ψ10 → ¬ψ20 )
Wk
{ϕ0 , ¬ψ1 } ` ψ10 → ¬ψ2 {ϕ0 , ¬ψ1 } ` ¬(ψ10 → ¬ψ20 )
¬i
{ϕ0 } ` ¬¬ψ1
¬e
{ϕ0 } ` ψ1
Moreover
Ax
{¬ψ20 } ` ¬ψ20
Wk Ax
{ϕ0 , ψ10 , ¬ψ20 } ` ¬ψ20 {ϕ0 } ` ϕ0
→i Wk
{ϕ0 , ¬ψ20 } ` ψ10 → ¬ψ20 {ϕ0 , ¬ψ20 } ` ϕ0
¬i
{ϕ0 } ` ¬¬ψ20
¬e
{ϕ0 } ` ψ20
induction
{ϕ0 } ` ψ2
22
• Suppose that ϕ = ψ1 ∨ ψ2 , so that ϕ0 = (¬ψ10 → ψ20 ).
– {ψ1 , ¬ψ10 } is clearly inconsistent by induction. By Lemma 0.3.6,
{ψ1 , ¬ψ10 } ` ψ20 and {ψ1 } ` ϕ0 .
On the other hand, {ψ2 , ¬ψ10 } ` ψ20 by induction, so {ψ2 } ` ϕ0 .
As ϕ = ψ1 ∨ ψ2 , combining both deduction trees and eliminating ∨,
we find {ϕ} ` ϕ0 . This was the easy part.
– For the converse we need an easy remark:
Ax
{ψ10 } ` ψ10
induction
{ψ10 } ` ψ1
Wk Ax
{¬ϕ, ψ10 } ` ψ1 {¬ϕ} ` ¬ϕ
∨i Wk
{¬ϕ, ψ10 } ` ψ1 ∨ ψ2 {¬ϕ, ψ10 } ` ¬ϕ
¬i
{¬ϕ} ` ¬ψ10
| {z }
(∗1 )
23
As all the rules in `0 are in `, one even has
` ψ10 → (ψ20 → . . . (ψn0 → ϕ0 ). . . )
Now one sees that
0
(ψ1 → (ψ2 → . . . (ψn → ϕ). . . )) is ψ10 → (ψ20 → . . . (ψn0 → ϕ0 ). . . )
So by Lemma 0.3.12, it follows
` ψ1 → (ψ2 → . . . (ψn → ϕ). . . )
In particular, {ψ1 , . . . , ψn } ` ϕ, and Σ ` ϕ.
We shall prove the converse implication next time.
End of Lecture 4.
Lecture 5 (Reduction to Two Connectives (2/2); Soundness)
We finish the proof of Corollary 0.3.14.
Proof of Corollary 0.3.14, completed. The remaining proof that Σ ` ϕ im-
plies Σ0 `0 ϕ0 is by induction on the length of the deduction.
• The cases of axiom and weakening are clear.
• Suppose the last step is for instance ¬i :
Σ ∪ {ψ} ` χ Σ ∪ {ψ} ` ¬χ
¬i
Σ ` ¬ψ
By induction, one has Σ0 ∪ {ψ 0 } `0 χ0 and Σ0 ∪ {ψ 0 } `0 ¬χ0 . As ¬i is in
`0 , it follows
Σ0 ∪ {ψ 0 } `0 χ0 Σ0 ∪ {ψ 0 } `0 ¬χ0
¬i
Σ0 `0 ¬ψ 0
And therefore Σ0 `0 ¬ψ 0 .
• It is clear that the cases of ¬e , →i , and →e are as trivial (since these rules
are in `0 ). We now turn to the interesting configurations.
• Suppose the last step is ∧i , that is
Σ ` ψ1 Σ ` ψ2
∧i
Σ ` ψ1 ∧ ψ2
We aim at proving Σ0 `0 (ψ1 ∧ ψ2 )0 ; recall that (ψ1 ∧ ψ2 )0 is ¬(ψ10 → ¬ψ20 ).
By induction, we know Σ0 `0 ψ10 and Σ0 ` ψ20 . Therefore
..
.. induction
Ax
Σ0 `0 ψ10 {ψ10 → ¬ψ20 } `0 ψ10 → ¬ψ20
Wk
Σ0 ∪ {ψ10 → ¬ψ20 } `0 ψ10 Σ0 ∪ {ψ10 → ¬ψ20 } `0 ψ10 → ¬ψ20
→e
Σ0 ∪ {ψ10 → ¬ψ20 } `0 ¬ψ20
| {z }
(∗)
24
and
..
.. induction
(∗) Σ0 `0 ψ20
Wk
Σ0 ∪ {ψ10 → ¬ψ20 } `0 ¬ψ20 Σ0 ∪ {ψ10 → ¬ψ20 } `0 ψ20
¬i
Σ0 `0 ¬(ψ10 → ¬ψ20 )
Σ ` ψ1 ∧ ψ2
∧e1
Σ ` ψ1
Σ ` ψ1 ∧ ψ2
∧e2
Σ ` ψ2
Σ ` ψ1
∨i1
Σ ` ψ1 ∨ ψ2
25
• Suppose the last step of the form:
Σ ` ψ1 ∨ ψ2 Σ ∪ {ψ1 } ` χ Σ ∪ {ψ2 } ` χ
∨e
Σ`χ
By induction, Σ0 `0 ¬ψ10 → ψ20 , Σ0 ∪ {ψ10 } `0 χ0 , and Σ0 ∪ {ψ20 } `0 χ0 . By
contraposition (Theorem 0.3.8, which still holds for `0 ), we easily find:
Σ0 ∪ {¬χ0 } `0 ¬ψ10 (∗1 ) and Σ0 ∪ {¬χ0 } `0 ¬ψ20 (∗2 )
Therefore
..
.. induction
Σ ` ¬ψ10 → ψ20
0 0
(∗1 )
Σ ∪ {¬χ0 } `0 ¬ψ10 → ψ20
0
Σ0 ∪ {¬χ0 } `0 ¬ψ10 (∗2 )
→e
0 0 0 0
Σ ∪ {¬χ } ` ψ2 Σ ∪ {¬χ0 } `0 ¬ψ20
0
¬i
Σ0 `0 ¬¬χ0
¬e
Σ0 `0 χ0
26
As Σ remains uncrossed, one has Σ ` ϕ → ψ.
Similarly,
Σ ∪ {ϕ} ` ψ Σ ∪ {ϕ} ` ¬ψ
¬i
Σ ` ¬ϕ
becomes:
Σ ϕ Σ ϕ Σ 6ϕ Σ 6ϕ
then
ψ ¬ψ ψ ¬ψ
¬ϕ
which means Σ ` ¬ϕ. Do not forget to cross out every relevant occurence of ϕ!
Example 0.3.16. As an example, we prove Excluded Middle (Theorem 0.3.11)
with this notation.
¬(ϕ ∨ ¬ϕ) ϕ
¬(ϕ ∨ ¬ϕ) ϕ ∨i
1. ϕ ∨ ¬ϕ 2.
¬(ϕ ∨ ¬ϕ) 6 ϕ
¬(ϕ ∨ ¬ϕ) 6 ϕ ∨i
∨i ϕ ∨ ¬ϕ
ϕ ∨ ¬ϕ ¬ϕ
¬i
¬i
¬ϕ ∨i
3. ϕ ∨ ¬ϕ 4.
¬(ϕ 6 ∨¬ϕ) 6 ϕ
¬(ϕ 6 ∨¬ϕ) 6 ϕ ∨i
∨i ϕ ∨ ¬ϕ
ϕ ∨ ¬ϕ ¬ϕ
¬i
¬i
¬ϕ ∨i
∨i ϕ ∨ ¬ϕ
ϕ ∨ ¬ϕ ¬i
¬i ¬¬(ϕ ∨ ¬ϕ)
¬¬(ϕ ∨ ¬ϕ) ¬e
5. ϕ ∨ ¬ϕ 6.
27
• The case of weakening is absolutely trivial.
• Suppose that the last step is by ¬i ; that is, Σ ` ¬ϕ because Σ ∪ {ϕ}
is inconsistent. By induction, there are no truth assignments satisfying
Σ ∪ {ϕ}. So any truth assignment satisfying Σ (if any) satisfies ¬ϕ. Thus
Σ |= ¬ϕ (notice that we did not claim satisfiability of Σ).
• Suppose that the last step is
Σ ` ¬¬ϕ
¬e
Σ`ϕ
End of Lecture 5.
28
Theorem 0.4.1 (completeness of propositional logic). Let Σ be a theory and
ϕ be a wff. If Σ |= ϕ, then Σ ` ϕ.
Lemma 0.4.2. The Completeness Theorem is equivalent to:
every consistent theory is satisfiable.
Proof . Suppose the Completeness Theorem. Let Σ be a consistent theory.
As Σ is consistent (Definition 0.3.5), there is a wff ϕ such that Σ 6` ϕ. By
completeness, this implies Σ 6|= ϕ. It follows from Lemma 0.2.10 that Σ ∪ {ϕ}
is satisfiable; let v be a truth assignment satisfying Σ ∪ {ϕ}. In particular, v
satisfies Σ!
Suppose that every consistent theory is satisfiable. Let Σ 6` ϕ. Then Σ∪{¬ϕ}
is consistent by Lemma 0.3.9. In particular, Σ ∪ {¬ϕ} is satisfiable. So there is
a truth assignment v satisfying Σ ∪ {¬ϕ}: v satisfies Σ but does not satisfy ϕ,
so Σ 6|= ϕ.
29
Proof . Recall that WFF is countable (Remark 0.1.3, for instance); enumerate
WFF = {ϕn : n ∈ N}. We shall extend Σ recursively. Let Σ0 = Σ. Assume Σn
has been constructed. By Lemma 0.4.5, Σn ∪ {ϕn } or Σn ∪ {¬ϕn } is consistent.
If Σn ∪ {ϕn } is consistent, let Σn+1 be it; otherwise let Σn+1 = Σn ∪ {¬ϕn }. So
the nth wff ϕn has been taken into account, and Σn+1 is still consistent.
Let Σ̂ = ∪n∈N Σn . Clearly Σ̂ is consistent, and for any wff ϕ, ϕ = ϕn for
some n, so ϕn or its negation is in Σn+1 ⊆ Σ̂. (It is also clear by consistency
that ϕ and ¬ϕ are not both in Σ̂.)
End of Lecture 6.
30
0.5 The Compactness Theorem
0.5.1 Compactness
Compactness is the general phenomenon of reduction of the infinite to the finite.
Definition 0.5.1 (finite satisfiability). Let Σ be a theory. Σ is finitely satisfiable
if for all finite Σ0 ⊆ Σ, Σ0 is satisfiable.
Theorem 0.5.2 (compactness of propositional logic). A theory is satisfiable iff
it is finitely satisfiable.
Proof . An implication is trivial: if Σ is satisfiable, then a truth assignment
satisfying it will also satisfy any finite subset.
Let Σ be a finitely satisfiable theory. Any finite fragment of Σ is satisfiable,
hence consistent by soundness (Theorem 0.3.17). So Σ is finitely consistent.
But consistency and finite consistency coincide, as consistency is a finite notion
(our deductions all have finite length). So Σ is consistent, therefore satisfiable
by completeness (Theorem 0.4.1 and Lemma 0.4.2).
Corollary 0.5.3. If Σ |= ϕ, then there is a finite Σ0 ⊆ Σ such that Σ0 |= ϕ.
Proof . By Lemma 0.2.10, Σ∪{¬ϕ} is not satisfiable; by compactness (Theorem
0.5.2) there is a finite subset Σ0 ⊆ Σ such that Σ0 ∪{¬ϕ} is not satisfiable. With
Lemma 0.2.10 again, this means Σ0 |= ϕ.
We now give an example of application.
Definition 0.5.4. A graph is k-colorable if there is a function c from the set
of vertices to {1, . . . , k} such that if {x, y} is an edge, then c(x) 6= c(y).
Recall that a subgraph of a graph is a subset of the set of vertices equipped
with all possible edges.
Theorem 0.5.5 (Erdös). A countable graph is k-colorable iff all of its finite
subgraphs are k-colorable.
Proof . Let V = {vn : n ∈ N} be an enumeration of the vertices. We add
sentence symbols Cv,i for v ∈ V and i ∈ {1, . . . , k} (their meanings will be:
c(v) = i, that is v bears the ith color).
Consider the theory made of:
• Cv,1 ∨ · · · ∨ Cv,k for each v ∈ V (meaning: each vertex has a color)
• ∧i6=j 6= Cv,i ∧ Cv,j for each v ∈ V (meaning: no vertex has two colors)
• ∧ki=1 6= Cv,i ∧ Cv0 ,i for adjacent vertices v, v 0 (meaning: no two adjacent
vertices bear the same color)
By assumption, the theory is finitely satisfiable. By compactness (Theorem
0.5.2), it is satisfiable: there exists a global coloring.
Still not impressed? Wait and see what we shall do with first-order logic. Ex. 0.14, 0.17
31
0.5.2 Applications to Decidability*
Compactness is very useful is the context of decidability, where one wants to
decide whether or not a given wff is a consequence of a theory. We consider this
problem. For the question to make sense, we need of course to have the data in
a form readable by a machine. This is approached by the following definition.
Definition 0.5.6 (decidable). A set is decidable if there is an algorithm which
can decide whether or not a given object belongs to it.
Example 0.5.7. The set of wff’s is decidable, as there is an algorithm which
determines whether a given formula is decidable or not.
Actually if one just wants to list all possible elements of the set, decidability
is a little too strong, and could be weakened as follow.
Definition 0.5.8 (semi-decidable). A set is semi-decidable if there is an algo-
rithmic way to write a list such that sooner or later, every element of the set
will appear on the list.
One also says effectively enumerable for semi-decidable. This does not assert
that there is an algorithm deciding whether or not an object is in the set; all you
can get is an algorithm which will answer “yes” if the object lies in it. Otherwise,
the algorithm doesn’t stop, as it scans through an infinite list without finding
an answer.
Example 0.5.9. Suppose that you are looking for Jorge Luis Borges in an
infinite phone directory.
• If the phone book is in alphabetical order, you start reading. If you reach
Byron without passing through Borges, then Jorge Luis Borges is not
in the phone book. If you reach Borges, Julio without passing through
Borges, Jorge, then Jorge Luis Borges is not in the phone book. Other-
wise you have reached Jorge Luis Borges. The phone book is decidable.
• But now suppose that the phone book is not in alphabetical order. The
only way to find out if Borges is in the phone book is to read through. If
you reach the name Borges, Jorge Luis, then he is in the phone book.
Otherwise, you cannot know whether he is not in the directory, or he is
but you haven’t found his name yet. The phone book is semi-decidable.
Remark 0.5.10. A set is decidable iff both it and its complement are semi-
decidable.
We now are ready to explain what a “readable” theory is.
Definition 0.5.11 (axiomatization). An axiomatization of a set of wff’s Σ is a
set Σ0 such that Σ0 |= Σ and Σ |= Σ0 .
Definition 0.5.12 (finitely, decidably, semi-decidably axiomatizable theory).
If there exists a finite, decidable, or semi-decidable axiomatization, the theory
is said finitely, decidably, semi-decidably axiomatizable, accordingly.
32
A typical use of Compactness is the following result.
Proposition 0.5.13. If Σ is finitely axiomatizable, then every axiomatization
of Σ contains a finite axiomatization.
Proof . Let Σ0 be a finite axiomatization of Σ; taking the conjunction, we may
assume that Σ0 = ϕ is a single formula.
Let Σ0 be any axiomatization of Σ. Then Σ0 ∪ {¬ϕ} is not consistent, as
Σ |= Σ |= ϕ. So Σ0 |= ϕ. By compactness, there is a finite subset of Σ0 , which
0
Proof . Run two instances of the algorithm of Corollary 0.5.14: one for ϕ and
one for ¬ϕ. One of them will sooner or later answer “yes”.
We do not insist further on decidability; our notion of algorithm is to remain
informal. Decidability and Computability are other branches of Mathematical
Logic, with very interesting results.
33
0.5.3 A Topological Proof*
Compactness phenomena first arose in the context of topology. Informally
speaking, a space is compact if whenever one can always find an element meet-
ing a finite number of requirements (lying in the intersection of finitely many
members), then there is an element meeting all requirements simultaneously.
And this is exactly what Theorem 0.5.2 says; it actually states the compact-
ness of a certain topological space. We therefore give an alternate proof of the
compactness theorem.
Definition 0.5.17 (finite intersection property). A family F has the finite
intersection property if the intersection of any finite number of sets in F is
nonempty.
Definition 0.5.18 (compactness). A topological space X is compact if for all
family of closed sets F with the finite intersection property, ∩F 6= ∅.
For instance, a finite space is always compact.
Notation 0.5.21.
• Let S = {F, T }N be the set of truth assignments.
• For every wff ϕ, let Oϕ = {v ∈ S : v(ϕ) = T }.
• Let τ be the topology generated by the Oϕ ’s.
Remark 0.5.23. This correspondence is not injective (as {ϕ} and {¬¬ϕ} define
the same closed set), but it is surjective since any closed set is an intersection
of clopen subsets.
Lemma 0.5.24. Let Σ be a theory and v be a truth assignment. Then v ∈ CΣ
iff v satisfies Σ.
34
Proof .
v ∈ CΣ iff for all ϕ ∈ Σ, v ∈ Oϕ
iff for all ϕ ∈ Σ, v(ϕ) = T
iff v satisfies Σ.
End of Lecture 7.
For instance (A1 → A2 ) is a wff of modal logic, which reads “It is
necessary that the possibility of A1 implies the necessity of A2 ”. Similarly, ¬ϕ
reads “ϕ is contingent”; ¬ ϕ reads “ϕ is impossible”.
0.6.1 Semantics*
The semantics is of course a little more complex than before. As mentioned, we
need a whole collection of worlds, accessible by imagination from each other.
Definition 0.6.2 (Kripke model). A Kripke model is a triple (W, R, v) where
35
• W is a non-empty collection of worlds;
• R is a binary relation on W ; w1 Rw2 means that w2 is conceivable in w1 ;
• v is a function from A × W to {T, F }.
• v(¬ϕ, w) = T if v(ϕ, w) = F
• v(ϕ → ψ, w) = T if v(ϕ, w) = F or v(ψ, w) = T
• v(ϕ, w) = T if for all world w0 such that wRw0 , one has v(ϕ, w0 ) = T
• v(ϕ, w) = T if there is a world w0 such that wRw0 and v(ϕ, w0 ) = T .
The meaning is clear: ϕ (“ϕ is necessary”) holds in the world w if for any
world accessible from w, ϕ holds. (An observer from w will then believe that ϕ
cannot ever fail.) On the other hand, ϕ (“ϕ is possible”) holds in w if there
is a conceivable world w0 in which ϕ holds: the observer of w will think that ϕ
could hold somewhere.
One can of course reduce the language to use the sole modal connective ;
clearly stands for ¬¬; we implicitely do so and forget about .
Definition 0.6.5 (satisfaction in modal logic). Σ |= ϕ if for all models (W, R, v)
such that Σ is satisfied in all worlds of W , ϕ is also satisfied in all worlds of W .
Example 0.6.6. One has {ϕ} |= ϕ.
• Kripke’s Rule:
Σ ` (ϕ → ψ)
K
Σ ` ϕ → ψ
36
• Necessitation Rule:
`ϕ
N
` ϕ
Caution! There are no assumptions in rule N , i.e. Σ must be the empty set.
Otherwise one finds
{ϕ} ` ϕ
N
{ϕ} ` ϕ
→i
` ϕ → ϕ
in other words, “everything which holds is necessary”, a clear insanity.
Recall however that {ϕ} |= ϕ; in particular, we shall not have “full” com-
pleteness, in the sense that Σ |= ϕ will be stronger than Σ ` ϕ. Completeness
will hold only for Σ = ∅.
Remark 0.6.7. Notice that if we have made specific requirements on the “ac-
cessibility” relation R (see Remark 0.6.3), we might need to add a couple of
extra rules.
Example 0.6.8. In Definition 0.6.2, we impose on the accessibility relation to
be reflexive. Then |= (ϕ) → ϕ. As there is no way to prove it, we admit it as
an axiom, or equivalently we add the following rule:
Σ ` ϕ
Σ`ϕ
Theorem 0.6.9 (soundness of modal logic). If Σ ` ϕ, then Σ |= ϕ.
Proof . A quick induction.
• The case of the rules of Propositional Logic is essentially as in Theorem
0.3.17. For
Σ ∪ {ϕ} ` ψ
→i
Σ`ϕ→ψ
you might need to introduce W1 = {w in W : v(ϕ, w) = T } and W2 =
{w in W : v(ϕ, w) = F }, and prove the claim piecewise.
• Suppose that the last step is
`ϕ
N
` ϕ
(Recall that the Necessitation rule has no assumptions).
Let (W, R, v) be any Kripke model. We show that for any world w of W ,
we have v(ϕ, w) = T .
By induction, |= ϕ, so our Kripke model satisfies ϕ: that is, for any w0 in
W , one has v(ϕ, w0 ) = T . In particular, for any w0 such that wRw0 , one
has v(ϕ, w0 ) = T . Hence v(ϕ, w) = T .
As w is arbitrary, the Kripke model satisfies ϕ. Since we have taken any
Kripke model, we deduce that |= ϕ.
37
• Now suppose that the last step is
Σ ` (ϕ → ψ)
K
Σ ` ϕ → ψ
0.6.3 Completeness*
Theorem 0.6.10 (completeness of modal logic). If |= ϕ, then ` ϕ.
Counter-example 0.6.11. This is not true with a non-empty set of assump-
tions, as {ϕ} |= ϕ, but {ϕ} `
6 ϕ.
Proof of Theorem 0.6.10. The proof will rely on the construction of a uni-
versal Kripke model (Wun , Run , vun ).
• Let Wun be the collection of all maximal consistent modal theories.
• Write wRun w0 iff {ϕ : ϕ ∈ w} ⊆ w0 .
38
• We now suppose vun (ϕ, w) = T ; we show that ϕ ∈ w. Consider Nw =
{χ : χ ∈ w}. We form the modal theory Θ = Nw ∪ {¬ψ}.
If Θ is consistent, then it can be extended to a maximal consistent modal
theory w0 . Hence Nw ⊆ w0 and by definition of the relation, wRun w0 . By
assumption, vun (ψ, w) = T , so vun (ψ, w0 ) = T , against ¬ψ ∈ w0 .
So Θ is not consistent. It follows Nw ` ψ. There are therefore χ1 , . . . , χn ∈
Nw such that {χ1 , . . . , χn } ` ψ. We find
{χ1 , . . . , χn } ` ψ
→i
` χ1 → (χ2 → (. . . (χn → ψ)...)
N
` (χ1 → (χ2 → (. . . (χn → ψ)...))
K
` χ1 → (χ2 → (. . . (χn → ψ)...)
Remark 0.6.12.
• Of course if we have made restrictions the accessibility relations (and clev-
erly added the corresponding deduction rules), we need to prove complete-
ness accordingly, that is we must show that the accessibility relation Run
on the universal model Wun has the desired properties.
This discussion reveals that the universal model WΣ satisfies the truth lemma
only if Σ is closed under . This yields the following generalization of Theorem
0.6.10.
Theorem 0.6.13 (completeness of modal logic revisited). Let Σ be a modal
theory which is closed under . If Σ |= ϕ, then Σ ` ϕ.
39
Theorem 0.6.14 (completeness of modal logic revisited). Let Σ be a modal
theory which is closed under . If Σ is consistent, then Σ is satisfiable.
Notice that the resulting compactness phenomenon does not require closure
under .
Corollary 0.6.15 (compactness of modal logic). Let Σ be a modal theory. Then
Σ is satisfiable iff it is finitely satisfiable.
Proof . Suppose Σ is finitely satisfiable; we prove that it is satisfiable (the other
implication being trivial).
Let Σ0 be the closure of Σ under : Σ0 is the smallest subset of WFFmod
which contains Σ and such that ϕ ∈ Σ0 =⇒ ϕ ∈ Σ0 . As for any ϕ ∈ WFFmod ,
{ϕ} ` ϕ, it is clear that Σ0 is finitely satisfiable.
By soundness of modal logic (Theorem 0.6.9), Σ0 is consistent. By complete-
ness (Theorem 0.6.14), Σ0 is satisfiable. As Σ ⊆ Σ0 , so is Σ.
However interesting epistemologically, one sees on the technical side that
modal logic is but an easy extension of propositional logic. We shall return to
more mathematical views.
End of Lecture 8.
Lecture 9 (ME1)
End of Lecture 9.
0.7 Exercises
When not otherwise specified, we work in classical logic (¬e allowed).
Exercise 0.1. Rewrite the following using only ¬ and →:
2. ϕ ∨ ψ 3. ϕ ↔ ψ
1. ϕ ∧ ψ
The idea is to show that our connectives are (intuitively speaking) redundant.
We shall show that they are formally redundant.
Exercise 0.2. The language has four symbols [, ], {, and }. The collection of
nice formulas is the smallest set of expressions such that:
• the empty/blank expression “” is a nice formula;
• if A, B are nice formulas, [A]{B} is a nice formula.
40
2. An expression is globally balanced if it is both bracket-balanced and brace-
balanced, and has as many [’s as {’s. Show that every nice formula is
globally balanced.
3. Show that no proper initial segment of a nice formula is globally balanced.
Deduce that no proper initial segment of a nice formula is a nice formula.
1. Σ1 = {A1 , A1 → A2 , A2 → A3 , A1 ∧ A2 , A2 ∨ ¬A3 }
2. Σ2 = {(A1 → A2 ) → A1 , ¬A1 }
3. Σ3 = {A3i ∧ A3i+1 → ¬A3i+2 : i ∈ N} ∪ {A2i+1 → ¬A2i+2 : i ∈ N}
Exercise 0.4. Let ϕ be a wff formula which uses as connectives only ∧ and
∨. Let v be the truth assignment such that v(An ) = T for all n. Show that
v(ϕ) = T .
Exercise 0.5.
1. Show that Σ ` ϕ → ψ iff Σ ∪ {ϕ} ` ψ.
Exercise 0.8.
1. Write a deduction of ` ϕ → ¬¬ϕ.
2. Show that ` ¬¬(ϕ ∨ ¬ϕ) (without using ¬e nor excluded middle).
41
• A0n = An for each sentence symbol;
• (¬ϕ)0 = (ϕ0 →⊥);
• (ϕ → ψ)0 = (ϕ0 → ψ 0 ).
We extend the notion of deduction ` to also include the ⊥ rules. Show
that Σ ` ϕ iff Σ0 ` ϕ0 .
As we know that other connectives may be coded in terms of ¬ and →, one sees
that all connectives can be coded by → and ⊥. A purely ⊥, → formulation of
¬e is
` ((ϕ →⊥) →⊥) → ϕ
Exercise 0.10 (Double negation, Excluded middle and Reductio ad Absur-
dum). Consider the rules:
Σ ` ¬¬ϕ Σ ∪ {¬ϕ} `⊥
¬e RAA EM
Σ`ϕ Σ`ϕ ` ϕ ∨ ¬ϕ
Show that all three are equivalent, i.e. taking one as a deduction rule allows to
establish the other two.
Exercise 0.11. Prove the soundness theorem by induction with all five con-
nectives.
Exercise 0.12. Let CONS be the property:
Every satisfiable theory is consistent.
Show that soundness is equivalent to CONS.
Exercise 0.13. Let FSAT be the property:
Every finite, consistent theory is satisfiable.
Show that completeness is equivalent to compactness and FSAT.
Exercise 0.14. Let X be a countable set and a partial ordering on X ((X, )
is a poset):
• ∀x xx
• ∀x ∀y xy∧y x→x=y
• ∀x ∀y ∀z x y ∧ y z → x z.
Show that there is a linear ordering ≤ on X extending , that is:
• ≤ is an ordering on X
• ≤ is linear: ∀x ∀y x≤y∨y ≤x
• ≤ extends : ∀x ∀y x y → x ≤ y.
42
Hint: When X is finite, an easy induction. When X is countable, use compact-
ness of propositional logic.
Exercise 0.15. A map is a countable partition of the plane satisfying certain
intuitive properties we do not wish to formalize. It is 4-colourable if there is
a function which assigns to each region a colour, such that no two adajacent
regions bear the same colour. Show that a countable map is 4-colourable iff
each of its finite submaps is.
Exercise 0.16. A graph is k-colorable if there is a function c from the set of
vertices to {1, . . . , k} such that if {x, y} is an edge, then c(x) 6= c(y). Show that
a countable graph is k-colorable iff all of its finite subgraphs are k-colorable.
Exercise 0.17.
1. Let {Sn : n ∈ N} be a collection of finite subsets of N with the following
property: for each finite F0 ⊆ N there exists A0 ⊆ N such that for all
n ∈ F0 , Card(A0 ∩ Sn ) = 1. Prove that there exists A ⊆ N such that for
all n in N, Card(A ∩ Sn ) = 1.
Hence the ¬ operation is not involutive in IL, but quickly becomes involutive.
Exercise 0.19 (Gödel’s translation). We work only with → and ⊥, as we know
(Exercise 0.9) that they are sufficient to encode all connectives.
We define inductively the Gödel translation of a wff
• A∗n = ¬¬An , An a sentence symbol (including ⊥)
• (ϕ → ψ)∗ = ϕ∗ → ψ ∗
It follows that Intuitionnistic Logic is more subtle than Classical Logic: it has
the ability to encode all of classical logic, and it perceives differences that CL
doesn’t. (It is weaker only when viewed from CL.)
43
Exercise 0.20 (some modal logic). We investigate further the accessibility
relation. A logic (|=, `) is tautology-complete if whenever |= ϕ, then ` ϕ.
1. Suppose that in any frame, the accessibility relation is reflexive (for w in
W , wRw), and that modal logic remains sound and tautology-complete.
Show that the following deduction is now valid:
` ϕ
`ϕ
Show that in the canonical model (constructed in the proof of the com-
pleteness theorem), the accessibility relation is now reflexive.
3. Deduce that if one considers only relfexive accessibility relations, and al-
lows the above rule, then modal logic remains tautology-complete.
44
Chapter 1
First-order Logic, also called Predicate Logic, is more expressive than Propo-
sitional Logic: it can mention elements. This requires extending the language;
in particular we shall need quantifiers. Up to an abbreviation, we use only one.
But this quantifier deeply affects the notion of deduction; we need new rules.
On the other hand, the meaning of first-order languages is also richer. We shall
introduce structures, which provide their semantic notion of entailment.
If soundness remains easy to show, completeness of first-order logic will be
a major and non-trivial result. Hence, though quite expressive, first-order logic
enjoys excellent properties.
In this chapter:
1.1 Syntax
First-order is essentially more expressive than Propositional Logic: we even-
tually introduce elements. So one will need quantification, but also specific
45
elements (constants), properties of elements (relations), and a way to build on
elements (functions).
46
• E1 6= E2 stands for ¬(= (E1 , E2 )).
This should save us precious amounts of time later, though one could in-
troduce five connectives (as we clumsily did for Propositional Logic before the
verifications of §0.2.3 and 0.3.3).
Remark 1.1.5. On the other hand, we have two quantifiers. By arguments
ressembling those of §0.2.3 and 0.3.3 (see Corollaries 0.2.18 and 0.3.14), we will
get rid of ∃, showing that it may be considered as a mere abbreviation. This
will be done in §1.4.2 (Corollaries 1.4.8 and 1.4.11).
Example 1.1.6.
• The empty language, or language of pure sets, contains the equality as its
only relation symbol.
• The language of orderings is Lord = {<}, with < a binary relation symbol
(we omit to mention =).
• The language of groups is Lgrps = {1, ·,−1 } where 1 is a constant symbol,
· is a binary function symbol, −1 is a unary function symbol.
• The language of rings is Lrings = {0, 1, +, −, ·} where 0 and 1 are constant
symbols, +, −, · are binary function symbols.
1.1.2 Terms
We have defined expressions, but only well-formed formulas are of interest. Yet
before we can define them, we need to explain which combinations of elements
are meaningful.
Definition 1.1.7 (term). The collection of L-terms is the smallest set such
that:
• every variable is a term;
• every constant symbol is a term;
• if t1 , . . . , tn are terms and f is a n-ary function symbol, then f (t1 , . . . , tn )
is a term.
Exercise 1.1
47
It is quite clear that a variable x occurs in a term t iff the symbol x appears
in the expression t.
Example 1.1.9.
• ·(1, ·(v1 , 1)) is a term of Lgrps , which is more conveniently written 1·(v1 ·1).
Only v1 occurs in it.
• 2 · v12 − 3 · v2 is a term of Lrings , as we have implicitely made the following
abbreviations: 2 stands for +(1, 1), 3 for +(1, +(1, 1)), and v12 for ·(v1 , v1 ).
The variables v1 and v2 occur in the term.
No meaning (and not much interest) so far! We start building on elements
to create well-formed formulas.
1.1.3 Formulas
Definition 1.1.10 (atomic formula). An atomic formula of L is an expression
of the form R(t1 , . . . , tn ), where R is an n-ary relation symbol of L and t1 , . . . , tn
are L-terms.
Example 1.1.11.
• v1 = v2 is an atomic formula of any first-order language (as equality is
always part of the language).
Theorem 1.1.14 (unique readability). There is only one way to read a wff.
Proof . Proceed exactly like for propositional logic, Theorem 0.1.9. This re-
quires a version of the balanced parenthesing lemma, Lemma 0.1.11.
48
Notation 1.1.15 (variables occuring in a formula). For an L-wff ϕ, we define
the set Var(ϕ) of variables occuring in ϕ:
• if ϕ is an atomic formula R(t1 , . . . , tn ), then Var(ϕ) = Var(t1 ) ∪ · · · ∪
Var(tn );
• if ϕ is ¬ψ, then Var(ϕ) = Var(ψ);
• if ϕ is ψ1 → ψ2 , then Var(ϕ) = Var(ψ1 ) ∪ Var(ψ2 );
• if ϕ is ∀x ψ or ∃x ψ, with x ∈ V, then Var(ϕ) = Var(ψ) ∪ {x}
Again, it is clear that a variable x occurs in a formula ϕ iff the symbol x
appears in the expression ϕ.
It is clear from Example 1.1.13 that when working syntactically we will have
to be extremely cautious with variables. This will give rise to several painful
technicalities, such as the notion of substitutability, to which the entire §1.3 is
unfortunately devoted.
Notice however that no one is stupid enough to make such mistakes at the
semantic level. So once we are done with the syntactic study of first-order
language, we shall happily forget the technical details.
Definition 1.1.16 (free or bound variable). Let ϕ be a wff and x a variable.
• if ϕ is atomic, then x occurs free in ϕ if it occurs in ϕ
• if ϕ is ¬ψ, then x occurs free in ϕ if it occurs free in ψ
• if ϕ is ψ1 → ψ2 , x occurs free in ϕ if it occurs free in ψ1 or in ψ2
• if ϕ is ∀y ψ or ∃y ψ, x occurs free in ϕ if x occurs free in ψ and x 6= y.
If x occurs in ϕ but is not free, it is said to occur bound in ϕ.
Example 1.1.17.
• In ∀v1 v1 = v2 , v1 is bound and v2 is free.
• The same holds of ∀v1 ∀v1 v1 = v2 (which is unfortunately a wff).
• In (∀v1 v2 = v3 ) → (v1 = v2 ), v1 , v2 and v3 occur free.
Hence free means: “we’re still missing something to know what it is about”.
Notation 1.1.18. For a formula ϕ, we let FreeVar(ϕ) denote the set of free
variables in ϕ.
Definition 1.1.19 (sentence). A sentence is a wff with no free variable.
Remark 1.1.20. In a sense, in propositional logic all (well-formed) formulas
are sentences, as there are no variables.
Definition 1.1.21 (theory). A theory is a set of sentences.
49
This definition is the most natural one (we do not want “unexplained vari-
ables”), but we shall need to be very careful.
• Some subtleties around the notion of consistency will arise in §1.4.
• We shall define Σ |= ϕ and Σ ` ϕ for sets of formulas, not of sentences, as
this will be useful later.
Example 1.1.22.
• The theory of infinite sets (in Lsets ) is
1.2 Semantics
Let us turn our attention to the meaning a first-order formula should convey.
This is a matter of interpreting the non-logical symbols (the signature of the
language), but also if necessary the free variables. This section is less clumsy
but more important than §1.1.
1.2.1 Structures
Definition 1.2.1 (L-structure). An L-structure M consists of:
• a non-empty base set M (called the base set or underlying set);
• for each constant symbol c of L, a specific element cM of M ;
• for each n-ary relation symbol R of L, a subset RM of M n ;
we require that =M must be the equality on M
50
• for each n-ary relation function f of L, a function f M from M n to M .
One also refers to M as the universe of the structure.
We have thus given a meaning to all symbols of L. All? No! We have
forgotten the meaning of the variables, which prevent the terms from having a
meaning, which prevent the formulas from having a meaning! So one must also
specify the value of the variables.
51
• if ϕ is ∀x ψ, then M |= ϕ[s] if it is the case that for any assignment s0
such that s and s0 agree on V \ {x}, M |= ψ[s0 ].
The idea is clear. The ∃ clause means that there is a value for x such that
ψ holds, the other variables being fixed as in s. The ∀ clause means that if one
freezes all variables y 6= x to s(y) but x assumes all possible values m ∈ M , ψ
holds in M.
Lemma 1.2.6 (univocity of satisfaction). Let M be an L-structure, ϕ an L-
formula, and s, s0 : V → M two assignments such that s| FreeVar ϕ = s0| FreeVar ϕ .
Then M |= ϕ[s] iff M |= ϕ[s0 ].
In particular, if ϕ is a sentence, then M |= ϕ[s] does not depend on s.
Proof . A clear induction.
We naturally write M |= Σ[s] if Σ is a set of wff’s such that for all ϕ ∈ Σ,
M |= ϕ[s].
If Σ is a set of sentences, then M |= Σ[s] does not depend on s; in this case
one says that M is a model of Σ. We do not insist on this terminology now, as
we shall be working mostly with sets of formulas instead of theories.
Definition 1.2.7 (satisfiability). Let Σ be a set of L-wff’s. Σ is satisfiable if
there is an L-structure M and an assignment s : V → M such that M |= Σ[s].
Pay attention to the following. We want our notion of consequence to deal
not only with sentences (which have no free variables), but also with formulas
with free variables. For instance we want to say that x = y does imply x + 1 =
y + 1, even though we have no idea of what x and y are. So we do need to
operate on all wff’s.
Definition 1.2.8 (semantic consequence). Let Σ be a set of formulas, ϕ a
formula. Σ |= ϕ if for any L-structure M and any assignment s : V → M , if
M |= Σ[s] then M |= ϕ[s].
Example 1.2.9.
• Let Tgrp be the theory of groups, ϕ the sentence ∀v1 v1 · v1 = 1, and ψ
the sentence ∀v1 ∀v2 v1 · v2 = v2 · v1 . Then Tgrp ∪ {ϕ} |= ψ.
• For n ∈ N, let ϕn be the sentence ∃v1 . . . ∃vn ∧i6=j ¬(vi = vj ). For each
n ≥ m, one has ϕn |= ϕm .
Lecture 12 (Substitutions)
1.3 Substitutions
We shall soon give deduction rules extending those of propositional logic. Before
that, we need more notions about the interplay of terms and formulas.
52
1.3.1 Substitutability
Given a formula ϕ and a variable x ∈ Var ϕ, we may want to replace x by a
term t. This will typically be the case in our proof theory, when we want to
deduce ϕ(t) from ∀x ϕ(x). But one must be extremely careful as the following
examples show.
Example 1.3.1.
• Consider the formula v1 = 1 and the term f (v1 ). Replacing, one finds
f (v1 ) = 1, and the replacement procedure stops after one iteration.
• Consider the formula ∀v1 v1 = v2 . If we replace v1 by a term which is not
a variable, the result no longer is a wff! (“∀1 1 = v1 ”)
We shall therefore substitute only free occurences. The problem is that
we cannot naively susbtitute all free occurences.
• If in ∃v1 v1 = v2 we replace v2 by v1 , we find ∃v1 v1 = v1 , which clearly
should mean something else.
• Similarly, if in ∃v1 v1 = v2 we replace v2 by any term involving v1 , we
alter the meaning.
These examples suggest that we may replace only certain free variables, but
not all. A problem arises when the replacement bounds a variable of t that was
free in ϕ before.
Definition 1.3.2 (term substitutable for a variable). A term t is substitutable
for a variable x in a formula ϕ if no variable of Var t becomes bound in ϕ when
one replaces every free occurence of x by t.
Example 1.3.3.
• Let ϕ1 be v1 = v2 and t be any term. Then t is substitutable for v1 in ϕ1 ;
it is also substitutable for v2 , and substitutable for v3 .
• Let ϕ2 be ∀v1 v1 = v1 . Then f (v1 ) is substitutable for v1 in ϕ2 , as there
are no free occurences of v1 . It is also substitutable for v2 , etc.
• Let ϕ3 be ∀v1 v1 = v2 . Then any term t is substitutable for v1 in ϕ3 . On
the other hand, t is substitutable for v2 in ϕ3 iff v1 does not occur in t.
For instance, if t = f (c, v1 ), then writing ∀v1 v1 = t would deeply alter
the structure of ϕ3 . If t = f (c, v2 ), or t = f (c, v3 ), then replacing v2 by t
is perfectly licit.
• Let ϕ4 be (∀v1 v1 = v2 ) → (v2 = v1 ) and t be a term. Then t is substi-
tutable for v1 in ϕ4 . On the other hand, t is substitutable for v2 in ϕ4 iff
v1 does not occur in t.
Notation 1.3.4. If t is substitutable for x in ϕ, one writes ϕ[t/x] for the
expression where every free occurence of x is replaced by t. (“/” is read “for”).
53
We refrain from writing ϕ[t/x] if t is not substitutable for x in ϕ.
Example 1.3.5. Consider Example 1.3.3 again.
• ϕ1 [t/v1 ] is t = v2 , whereas ϕ1 [t/v3 ] is of course v1 = v2 .
• ϕ2 [f (v1 )/v1 ] is ∀v1 v1 = v1 . So is ϕ2 [f (v1 )/v2 ].
• ϕ3 [f (c, v1 )/v1 ] is ∀v1 v1 = v2 .
ϕ3 [f (c, v3 )/v2 ] is ∀v1 v1 = f (c, v3 ).
“ϕ3 [f (c, v1 )/v2 ]” is of course not allowed (not substitutable), as the v1
from t would become bound.
• ϕ4 [f (c, v1 )/v1 ] is (∀v1 v1 = v2 ) → (v2 = f (c, v1 )).
ϕ4 [f (c, v3 )/v2 ] is (∀v1 v1 = f (c, v3 )) → (f (c, v3 ) = v1 ).
“ϕ4 [f (c, v1 )/v2 ]” is not allowed.
Bear in mind that we replace only free occurences!
54
If x 6= y but y ∈ Var(t), then something must be done. We let z be any
(or better: the first!) variable not in Var(ψ̃) ∪ Var(t) ∪ {x}. Notice that
z is then substitutable for y in ψ̃ (as it doesn’t appear in it); so ψ̃[z/y]
makes sense. Let ϕ̃ be ∀z ψ̃[z/y]. We claim that t is substitutable for x
in ϕ̃: this is clearly the case, as it is true in ψ̃, and z does not occur in t.
• We use the same method to deal with ∃y ψ.
Notice that if V has been effectively enumerated, then taking the first vari-
able not occuring in Var(ψ̃) ∪ Var(t) ∪ {x} makes ϕ̃ well-defined, and we have
an effective algorithm.
Example 1.3.7. Example 1.3.3 again.
• The easiest is to take ϕ̃1 for ϕ1 . One can substitute any term for any
variable in ϕ1 .
• Similarly, ϕ̃2 can be ϕ2 , as any term is substitutable for any variable in
ϕ2 .
• t is substitutable for v1 in ϕ3 , so with respect to v1 , ϕ̃3 could be ϕ3 .
If v1 6∈ Var(t), then t is substitutable for v2 in ϕ3 ; ϕ̃3 can still be ϕ3 .
Now if v1 ∈ Var(t) and we want to susbtitute t for v1 in ϕ3 , something
must be done. If for instance t is f (v1 ), then “∀v3 v3 = v2 ” will do as
ϕ̃3 . (In this case, ϕ̃3 [t/v1 ] is “∀v3 v3 = f (v1 )”.) On the other hand if t is
g(v1 , v2 , v3 ), then the ϕ̃3 given by our algorithm will be ∀v4 v4 = v2 (and
ϕ̃3 [t/v1 ] will now be ∀v4 v4 = g(v1 , v2 , v3 )).
Notice that ϕ̃3 does depend on x and t!
• We want to be able to substitute t for v2 in ϕ4 ; a problem occurs only if
v1 ∈ Var(t), which we assume. Let z be a variable not occuring in t, nor
equal to v2 . Then “(∀z z = v2 ) → (v2 = v1 )” will do as ϕ̃4 . (Of course
this ϕ̃4 only fits a certain desired substitution, as z depends on t.)
If we were not interested in the proof theory of first-order logic, we would
work permanently “up to renaming”. Unfortunately we will have to be a little
careful.
55
Proof . Induction on τ . For clarity, let τ̌ denote τ [t/x]. We want to prove that
š(τ ) = s(τ̌ ).
• This is clear if τ is a constant, or a variable not equal to x.
• If τ is x, then τ̌ is t, and š(τ ) = š(x) = s(t) = s(τ̌ ).
• If τ is f (τ1 , . . . , τn ), then š(τ ) = š(f (τ1 , . . . , τn )) = f M (š(τ1 ), . . . , š(τn )) =
f M (s(τ̌1 ), . . . , s(τ̌n )) = s(f (τ̌1 , . . . , τ̌n )) = s(τ̌ ).
Proposition 1.3.9. Let ϕ be a formula and t a term substitutable for a variable
x in ϕ. Let M be an L-structure and s : V → M an assignment. Let š be
the assignement which agrees with s on V \ {x} but on š(x) = s(t). Then
M |= (ϕ[t/x])[s] iff M |= ϕ[š].
Proof . Induction on ϕ. For clarity we let ϕ̌ denote ϕ[t/x]. Hence we want to
prove M |= ϕ̌[s] iff M |= ϕ[š].
• Suppose that ϕ is atomic, say R(τ1 , . . . , τn ). Then ϕ̌ is the formula
R(τ1 [t/x], . . . , τn [t/x]), which we naturally denote R(τ̌1 , . . . , τ̌n ). Now us-
ing Lemma 1.3.8,
M |= ϕ̌[s] iff (s(τ̌1 ), . . . , s(τ̌n )) ∈ RM
iff (š(τ1 ), . . . , š(τn )) ∈ RM
iff M |= ϕ[š]
56
– Now suppose M |= ϕ[š]. We want to show that M |= ϕ̌[s], that is
M |= (∀y ψ̌)[s]. So let s0 be any assignement agreeing with s except
on y. Let š0 agree with š except on y, and such that š0 (y) = s0 (y).
Notice that š0 (x) = š(x) = s(t) = s0 (t), as y 6∈ Var(t), and s and s0
agree except on y.
We know M |= ϕ[š], so M |= (∀y ψ)[š]. In particular, M |= ψ[š0 ].
But š0 and s0 agree everywhere but on x, and š0 (x) = s0 (t). So
by induction, M |= ψ̌[s0 ]. As s0 was arbitrarily obtained from s
by changing its value on y, we find M |= (∀y ψ̌)[s], and therefore
M |= ϕ̌[s].
• The case of ∃y ψ is similar.
Counter-example 1.3.10. It is extremely important to have t substitutable
for x in ϕ. Consider the case where ϕ is ∀v2 v1 = v2 and t is v2 . Clearly t is not
substitutable for v1 in ϕ. We let ϕ̌ denote the sentence ∀v2 v2 = v2 , resulting
from an illegitimate replacement.
It is the case that for any L-structure M and assignment s : V → M one will
have M |= ϕ̌[s]. But it will not be the case in general that M |= ϕ[š]. Reasoning
as in the proof of Proposition 1.3.9, one does have š0 (x) = š(x) = s(t), but now
y occurs in t, and there is no reason why s(t) = s(y) and s0 (y) should agree!
End of Lecture 12.
Lecture 13 (Deductions; Simplifying the Language)
• ∀ adds deduction rules. One may infer ∀x ϕ from ϕ if x does not appear in
th assumption. On the other hand, using a universal quantifier amounts
to substituting to a special case.
57
• ∃ adds dual deduction rules. One shows an existential by providing an
example, and uses one by introducing a witness.
Exercise 1.9
Of course we shall prove that these rules may be regarded as short-cuts,
provided ∃ is considered as an abbreviation for ¬∀¬ (this will be done in
§1.4.2).
One could now finish the proof of the renaming algorithm (Theorem 1.3.6)
by showing that {ϕ} ` ϕ̃ and {ϕ̃} ` ϕ. Exercise 1.14
` ∀x ∀y (x = y → y = x)
58
• for atomic ϕ, ϕ0 is ϕ;
• (¬ϕ)0 is (¬ϕ0 );
• (ϕ → ψ)0 is (ϕ0 → ψ 0 );
59
• Notice that x is substitutable for x in ψ 0 . Using induction, one finds
Ax,Wk
{∃x ψ, ψ, ∀x ¬ψ 0 } ` ψ
.. Ax,Wk
.. induction {∃x ψ, ψ, ∀x ¬ψ 0 } ` ∀x ¬ψ 0
∀e
{∃x ψ, ψ, ∀x ¬ψ 0 } ` ψ 0 {∃x ψ, ψ, ∀x ¬ψ 0 } ` ¬ψ 0
¬i
{∃x ψ, ψ} ` ¬∀x ¬ψ 0
| {z }
(∗)
so
{∃x ψ} ` ∃x ψ (∗)
∃e
{ϕ} ` ϕ0
Ax,Wk
{¬∃x ψ, ψ 0 } ` ψ 0
..
.. induction
{¬∃x ψ, ψ 0 } ` ψ
∃i Ax,Wk
{¬∃x ψ, ψ 0 } ` ∃x ψ {¬∃x ψ, ψ 0 } ` ¬∃x ψ
¬i
{¬∃x ψ} ` ¬ψ 0
| {z }
(∗)
{¬∃x ψ} ` ¬ψ 0
∀i
{¬∃x ψ} ` ∀x ¬ψ 0
60
• Suppose that the last step is:
Σ ` ψ[t/x]
∃i
Σ ` ∃x ψ
where t is substitutable for x in ψ. Then
.. Ax
.. induction {∀x ¬ψ 0 } `0 ∀x ¬ψ 0
Wk
Σ0 `0 ψ 0 [t/x] Σ0 ∪ {∀x ¬ψ 0 } `0 ∀x ¬ψ 0
Wk ∀e
Σ0 ∪ {∀x ¬ψ 0 } `0 ψ 0 [t/x] Σ0 ∪ {∀x ¬ψ 0 } `0 ¬ψ 0 [t/x]
¬i
Σ0 `0 ¬∀x ¬ψ 0
| {z }
=ϕ0
so Σ0 `0 ϕ0 .
• Now suppose that the last step is:
Σ ` ∃x ψ Σ ∪ {ψ} ` ϕ
∃e
Σ`ϕ
where x does not occur free in Σ ∪ {ϕ}. By induction, Σ0 ∪ {ψ 0 } `0 ϕ0 ;
contraposing (Theorem 0.3.8), Σ0 ∪ {¬ϕ0 } `0 ¬ψ 0 .
Now
Σ0 ∪ {¬ϕ0 } `0 ¬ψ 0
∀i
Σ0 ∪ {¬ϕ0 } `0 ∀x ¬ψ 0
since x does not occur free in Σ0 ∪ {ϕ0 }.
By induction again, Σ0 `0 ¬∀x ¬ψ 0 . So Σ0 ∪ {¬ϕ0 } is inconsistent, and
Σ0 `0 ϕ0 by Reductio ad Absurdum (Theorem 0.3.10).
Hence for us, an existential statement may be proved by showing that a
universal statement does not hold. Intuitively, this is disputable: one has never
given an example! For instance, in intuitionistic logic where one does not allow
¬e as a deduction rule, ∃ is not equivalent to ¬∀¬. Exercise 1.15
1.4.3 Soundness
Theorem 1.4.12 (soundness of first-order logic). If Σ ` ϕ, then Σ |= ϕ.
Proof . By induction on the deduction. Notice that we are extending propo-
sitional logic, so weakening, negation, and implication have been dealt with
in the proof of the propositional logic version (Theorem 0.3.17). Axioms of
equality are easily dealt with. It remains to consider the case of quantification.
By Corollaries 1.4.8 and 1.4.11, only ∀ need be examined: ∃ is now a mere
abbreviation.
61
• Suppose that the last step of the deduction was:
Σ`ϕ
∀i
Σ ` ∀x ϕ
1.5 Completeness
Soundness (Theorem 1.4.12) was the easy part. We now show the analog of
Theorem 0.4.1 for first-order logic; informally speaking, it expresses that “if it
is true, then there is a deduction of it" (in first-order classical logic for natural
deduction).
This theorem is due to Kurt Gödel (in his PhD).
Theorem 1.5.1 (completeness of first-order logic). If Σ |= ϕ, then Σ ` ϕ.
We adopt the same strategy as for the case of propositional logic (Theorem
0.4.1). It suffices to show the following:
Theorem 1.5.2 (completeness of first-order logic, equivalent version). If Σ |=
ϕ, then Σ ` ϕ.
62
Theorems 1.5.1 and 1.5.2 are equivalent (see Lemma 0.4.2).
Remark 1.5.3. Throughout we shall be working with sets of formulas instead
of theories (which are sets of sentences: Definition 1.1.21). Even if you have
a feeling that it would suffice to show Theorem 1.5.2 for theories, there will
be a point in the proof (the fairly technical Lemma 1.5.28 below) in which all
formulas must be taken into account.
Such a subtlety was impossible to imagine in propositional logic, where in a
sense all formulas are sentences!
Remark 1.5.4. The consistent set we start with may be assumed to contain
the axioms of equality. This remark is essential; see Remark 1.4.1.
Such a formula means: “If ϕ is not always true, it is because of cx,ϕ precisely”.
But this will require extending the language with new constants (§1.5.1),
and extending the collection of formulas to force this phenomenon (§1.5.2).
This strategy of adding witnesses is due to Leon Henkin.
Remark 1.5.5. We shall only treat the countable case, but some remarks will
indicate how to generalize to uncountable languages.
Let Σ be a consistent set of L-formulas containing the axioms of equality.
The goal is to find a nice extension of Σ in a language which provides witnesses.
63
1.5.1 Expanding the Language
We want to introduce witnesses for formulas of the form ¬∀x ¬ϕ. This raises
two (minor) issues.
• When we add new constants, we expand the language, so we also create
new formulas, and we need even more witnesses!
• When we add new constants, we expand the class of terms, so we also
expand the equality axioms and the ∀ rules. Will Σ still be consistent?
We deal with the first issue. The intuitive idea is that for any formula of
the form ∀x ϕ, we add a new constant symbol cx,ϕ . Of course when we expand
the language, we add new formulas! It thus looks like we need to repeat the
construction; fortunately the cardinal remains the same.
Lemma 1.5.6. There is a countable extension L0 of L by a set C 0 of new
constant symbols cx,ϕ , such that C 0 is in bijection with the set of pairs (x, ϕ)
where x ∈ V and ϕ is a L0 -formula.
Proof . Let L0 = L and WFF0 = WFF(L0 ). For each pair (x, ϕ) ∈ V × WFF0 ,
we add a new constant symbol cx,ϕ , getting the language L1 .
We repeat the construction with formulas of WFF1 = WFF(L1 ), and intro-
duce new constant symbols, which yields L2 . Let L0 = ∪n∈N Ln . Clearly L0 has
the desired property.
Notice that at each step Card Ln remains countable, so Card L0 = Card L.
We now deal with the second possible issue - why does Σ remain consistent in
L0 ? The intuitive reason is clear: if from Σ one can deduce a contradiction in L0 ,
then the deduction has used new constant symbols; L says nothing about these,
so they behave like free variables. This is expressed by the following theorem.
In the statement below, ϕ[z/c] denotes brute replacement of all occurences of
the constant c by the variable z (this creates no confusion, as there was certainly
no quantification over the constant c).
Theorem 1.5.7 (generalized generalisation). Let c be a constant symbol occur-
ing in ϕ, and suppose Σ ` ϕ where c does not occur in Σ. Then there is a
variable z not occuring in ϕ such that Σ ` ∀z ϕ[z/c].
Example 1.5.8. If Σ ` c = 0 and Σ does not mention c, then Σ ` ∀x x = 0.
Proof of Theorem 1.5.7. We write D : Σ ` ϕ to say that D is a deduction
(our usual deductive tree). Notice that there is a finite subset Σ0 ⊆ Σ such that
D : Σ0 ` ϕ.
Let z be a variable not occuring in D nor Σ0 (which are finite, but we have
countably many variables!) We replace c by z in each node of D, forming a tree
D[z/c]. We show that D[z/c] : Σ0 ` ϕ[z/c].
• If we use an axiom of equality, then substituting z for c remains an axiom
of equality!
64
• If the step is for instance of the form:
Σ0 ` ψ Σ0 ` ψ → χ
→e
Σ`χ
Σ0 ` ψ[z/c] Σ0 ` (ψ → χ)[z/c]
→e
Σ0 ` χ[z/c]
where x does not occur free in Σ0 . Notice that by choice of z, (∀x ψ)[z/c]
is ∀x (ψ[z/c]). It follows
..
.. induction
Σ0 ` ψ[z/c]
∀i
Σ0 ` (∀x ψ)[z/c]
Remark 1.5.9. Σ0 appears just to make sure that we can provide a new variable
z: perhaps Σ did mention all variable names... but we used only finitely many
in the deduction anyway.
Corollary 1.5.10. Σ with the new equality axioms remains consistent in L0 .
65
Proof . We shall prove a little more: write Σ `0 ϕ if there is a proof using
deduction rules associated to L0 , and Σ ` ϕ if the deduction is carried in L. We
claim that if Σ, ϕ do not use new constant symbols, then Σ `0 ϕ implies Σ ` ϕ.
This will clearly establish L0 -consistency of Σ.
So we suppose Σ `0 ϕ, and we assume Σ finite. Working inductively, for
every new constant symbol ci ∈ C 0 which appears in the deduction, there is
by Theorem 1.5.7 a variable xi such that Σ ` ∀xi ϕ[xi /ci ]; in particular Σ `
ϕ[xi /ci ]. As the ci ’s don’t appear in ϕ, we thus have Σ ` ϕ.
Yes! We have found a reasonable language to work in: it has enough wit-
nesses, and Σ remains consistent.
Remark 1.5.11. The construction of L0 immediately generalizes to uncount-
able L (the construction remains indexed by N; at each stage, Card Ln =
Card L). Corollary 1.5.10 is still true.
where cxn ,ϕn is the first element of C 0 which does not occur in {ϕ0 , . . . , ϕn } ∪
{η0 , . . . , ηn−1 }. For simplicity, we write cn instead of cxn ,ϕn (there is no risk of
confusion with the original constants).
ηn means: “if ϕn is not true for all xn ’s, it is because of...” and points to a
specific constant symbol, which has never been mentioned before.
Notation 1.5.13. Let Σ0 = Σ ∪ {ηn : n ∈ N}.
We have expanded our consistent set. Is it still consistent?
Lemma 1.5.14. Σ0 is a consistent set of L0 -formulas.
Proof . Let Σn = Σ ∪ {ηk : k ≤ n}. We know that Σ0 = Σ is consistent.
Suppose that Σn is; we show that Σn+1 is consistent again.
Otherwise, Σn ` ¬ηn+1 . Recall that ηn+1 stands for
66
Minimal work shows Σn ` ¬∀xn+1 ϕn+1 and Σn ` ϕn+1 [cn+1 /xn+1 ]. But
by construction (Notation 1.5.12), cn+1 does not occur in Σn . So generalizing
on the constant (Theorem 1.5.7), we may replace cn+1 by xn+1 and quantify.
We find Σn ` ∀xn+1 ϕn+1 , which is a contradiction to its consistency.
Σ0 is the basis of our inductive process; it is a consistent set of L0 -formulas.
We now maximize the set of requirements (see §0.4.1). Unfortunately, as ob-
served at the beginning of the proof of completeness (see Remark 1.5.3), we
cannot limit ourselves to sentences, but must take all possible formulas into
account.
Definition 1.5.15 (V-complete; see Definition 0.4.3). A consistent set of first-
order formulas Σ̂ is V-complete if for any formula ϕ, either ϕ or ¬ϕ is in Σ̂.
Caution! This is way stronger than what complete ought to be. A V-complete
(for: variable-complete) set of formulas even prescribes the behaviour of the
variables (v1 = v2 is in it, or v1 6= v2 is in it). The good notion, when working
with theories, is that of completeness (Definition 1.6.2 below), which is only
about sentences.
Remark 1.5.16 (see Remark 0.4.4). If Σ̂ is V-complete, then for any wff ϕ one
has Σ̂ ` ϕ iff ϕ ∈ Σ̂.
We now make our way to a V-complete set of formulas containing the Henkin
axioms.
Lemma 1.5.17 (see Lemma 0.4.5). If Σ is consistent and ϕ is a wff, then
Σ ∪ {ϕ} or Σ ∪ {¬ϕ} is consistent.
Proof . Very much like Lemma 0.4.5.
Lemma 1.5.18 (see Lemma 0.4.6). Let Σ be a consistent set of wff’s. Then
there is a V-complete set Σ̂ containing Σ and the Henkin axioms.
Proof . Like Lemma 0.4.6.
67
1.5.3 Finding a Structure (and Assignment)
Recall that from a set Σ, we built a maximal, consistent set of L0 -formulas
Σ̂ which contains all axioms of L0 , for some expansion L0 of L by new con-
stant symbols, and the Henkin axioms (Notation 1.5.12), which explain their
behaviours.
We are very close to satisfying Σ̂.
Notation 1.5.20. Let T 0 be the set of terms of L0 .
This looks like a very clever base set, but here is a minor complication. If
in the language there are a constant symbol c and a unary function symbol f ,
then c, f (c), etc. are terms. But perhaps the formula f (f (c)) = c is in Σ̂, in
which case the terms f (f (c)) and c are required to code the same element.
To force this, we identify pairs of terms which are believed to be equal by
the set of conditions.
Notation 1.5.21. Let ∼ be the relation on T 0 :
t ∼ t0 if Σ̂ ` t = t0
68
We have an L-structure, and now need an assignment of the variables; but
as they are terms, they already have been taken care of.
Notation 1.5.26. Let s : V → M map x to [x].
Lemma 1.5.27. For any L0 -term t ∈ T 0 , s(t) = [t].
Proof . Clear for a constant and a variable; clear by induction.
Lemma 1.5.28 (truth lemma; see Lemma 0.4.7). For any L0 -formula ϕ, one
has M |= ϕ[s] iff ϕ ∈ Σ̂.
Proof . Induction on ϕ. Bear in mind that by V-completeness of Σ̂, for any
formula, ϕ ∈ Σ̂ iff Σ̂ ` ϕ!
• Suppose ϕ is t1 = t2 . Then
• Suppose ϕ is ψ → χ. Then
69
– Assume M 6|= ϕ[s]; we want to show that ϕ 6∈ Σ̂. By assumption,
there is an assignment s0 which agrees with s except on xn and such
that M |= (¬ϕn )[s0 ].
By construction of M , s0 (xn ) is some class [t] for a term t. We would
like to use Proposition 1.3.9 and say that “M |= (¬ϕn [t/xn ])[s]”; the
issue is that perhaps t is not sustitutable for xn in ϕn , so the above
does not make sense. We need to rename a bit.
Let ϕ̃n be a formula in which t is substitutable for xn and such that
{ϕn } ` ϕ̃n and {ϕ̃n } ` ϕn (Theorem 1.3.6). Then M |= (¬ϕ̃n )[s0 ],
s0 (xn ) = [t] = s(t) (Lemma 1.5.27), and t is substitutable for xn
in ϕ̃n . The use of Proposition 1.3.9 is now legitimate, and we find
M |= (¬ϕ̃n [t/xn ])[s].
By induction, ϕ̃n [t/xn ] 6∈ Σ̂. By consistency (and as t is substituable
for xn in ϕ̃n ), ∀xn ϕ̃n 6∈ Σ̂ either. We are done with substitutions;
as ϕ̃n and ϕn prove each other, we find ∀xn ϕn 6∈ Σ̂. So ϕ 6∈ Σ̂.
The reader should have another look at remark 1.5.3; it is now clear that we
had to work with all formulas, and not only sentences.
Proof of Gödel’s Completeness Theorem for First-Order Logic. Fix a
first-order language L and a consistent set of L-formulas Σ. We find an L-
structure and an assignment satisfying Σ.
We first expand L to a language L0 having many witnesses (Lemma 1.5.6).
We then extend Σ to a V-complete set of L0 -formulas Σ̂, which describes the
behavior of the witnesses (Lemma 1.5.18).
We construct an L0 -structure M as in Notation 1.5.24 and an assignment s
as in Notation 1.5.26. By Lemma 1.5.28, M |= ϕ[s] for all ϕ ∈ Σ0 . In particular,
M |= Σ[s]: Σ is satisfiable.
End of Lecture 15.
70
We want to generalize Corollary 0.5.16 to first-order logic; this requires defin-
ing some maximality condition for a theory. V-completeness (Definition 1.5.15,
an ad hoc notion for the proof of the completeness theorem) is obviously too
strong, as one should care only for sentences. The good notion is the following.
Definition 1.6.2 (complete). A first-order theory Σ is complete if for any
sentence ϕ, either Σ ` ϕ or Σ ` ¬ϕ.
Notice that Σ is complete iff maximal as a consistent theory. This is more
understandable if one closes Σ under consequence, in which case this definition
is the same as the one we gave when proving completeness of propositional logic,
Definition 0.4.3.
Caution! Let us insist that as opposed to V-completeness, only sentences are
taken into account. Completeness is much weaker than V-completeness, but is
the only relevant notion when working with theories.
1.6.2 Compactness
Theorem 1.6.4 (compactness of first-order logic). Let Σ be a set of first-order
formulas. Then Σ is satisfiable iff Σ is finitely satisfiable.
Proof . If Σ is finitely satisfiable, it is consistent. As it is consistent, it is
satisfiable by the equivalent form of completeness.
Caution! Even if you have a clear idea where to satisfy finite fragments of Σ,
you may be surprised by a structure satisfying Σ!
Example 1.6.5. Consider an infinite set A, and let the language contain con-
stants ca (a ∈ A). The set of formulas Σ = {v1 6= ca : a ∈ A} is clearly finitely
satisfiable in A, as it suffices to take an assignment with s(v1 ) not one of the
finitely many mentioned a’s. But of course no assignment to A will satisfy all
of Σ simultaneously.
Corollary 1.6.6. If Σ |= ϕ, there is a finite subset Σ0 ⊆ Σ such that Σ0 |= ϕ.
Proof . Like Corollary 0.5.3.
One may want to have another look at the notion of finite axiomatizability
(Definition 0.5.12).
Example 1.6.7. The theory of an infinite set is not finitely axiomatisable.
71
Verification: Suppose it is. Then by Proposition 0.5.13 the natural axiomati-
zation {∃x1 . . . ∃xn ∧i6=j xi 6= xj : n ∈ N} too contains a finite axiomatisation,
which reduces to a single sentence {∃x1 . . . ∃xn ∧i6=j xi 6= xj }. The latter is
satisfied by any possibly finite set with at least n elements, a contradiction. ♦
Before moving to an application, let us comment on a historical misunder-
standing.
72
manner. We shall form an ad hoc theory, which will force the structure to
behave.
Consider the language L = {0, 1, <, +, −, ·, |.|} and the ordered field R. We
add new constants {cr : r ∈ R} (which will stand for elements of R), finding the
language LR , and one more new constant α; let L0 = LR ∪ {α} be the resulting
first-order language. Let Σ be the set containing the following:
• all LR -sentences which hold true in R;
• the formulas {0 < α < cr : r ∈ R>0 }
(One cannot formally quantify over r in the latter: the cr ’s are elements
of the language, but their collection is not expressible in the language!)
The theory Σ thus states that the cr ’s have the same relations as the real
numbers they stand for (in particular, some sentences express that we are talking
about an ordered field); the extra formulas say that α is a positive element
smaller than any positive real number.
73
• o = {x ∈ R∗ : for all r ∈ R>0 : |x| < r}.
b is the set of bounded elements (“not infinitely large”); o is the set of
infinitesimals (“infinitely close to 0”). Of course o ∩ R = {0}. If you want to
have a better picture of b and o, you may consider that the real line R embeds
into the non-standard line R∗ . b is the portion of the non-standard line which
is not incommensurably longer than R. o is the thickness around ◦ .
It goes without saying that though nicely defined, these subsets of R∗ are
not (first-order)-definable: the definitions are not expressible in our language
L0 .
Lemma 1.6.12. b is a ring, and o is an ideal of b.
Proof . Easy computations.
We shall now define an isomorphism b/o ' R. The construction is fairly
interesting. Following the same mental picture as before, when one consider
only the commensurable part and factors out the thickness, one retrieves the
standard line R.
Definition 1.6.13 (standard part). For any non-negative x ∈ b, let st x be the
least upper bound of {r ∈ R : r < x}. (We define st x for negative x ∈ b by
st x = − st(−x)).
This makes sense, as R is Dedekind-complete. Let x ∈ b be non-negative;
then the subset A = {r ∈ R : r < x} of R is non-empty and bounded above
since x ∈ b. So A has a least upper bound in R; st is well-defined. A little more
basic algebra reveals:
Lemma 1.6.14. st is a ring-homomorphism with kernel o.
Now as st is clearly onto R (it is the identity on R), we deduce that st
induces an isomorphism b/o ' R. In particular, any element x ∈ b can be
written in a unique way as the sum of a real number st x (its standard part),
and an infinitesimal ε ∈ o. (This is of course not true for elements of R∗ \ b,
which are infinitely big.)
One can then start doing analysis in a very elegant fashion, whith arguments
such as: f is continuous at real a ∈ R iff for any infinitesimal ε, f (a + ε) − f (a)
is infinitesimal.
74
• ∅ 6∈ F;
• if A ∈ F and A ⊆ B, then B ∈ F;
• if A, B ∈ F, then A ∩ B ∈ F.
Example 1.7.2. In a topological space, the collection of neighborhoods of a
given point forms a filter.
Remark 1.7.3. Let B be a family of subsets of X having the finite intersection
property (Definition 0.5.17). Then there is a filter F such that B ⊆ F.
Verification: Let F = {Y ⊆ X : there exist B1 , . . . , Bn in B such that B1 ∩
· · · ∩ Bn ⊆ Y }, i.e. F is the smallest family of sets containing every finite
intersection of members of B. This clearly meets the conditions defining a filter.
Notice that ∅ 6∈ F by the finite intersection property. ♦
Recall that Y cofinite in X means that X \ Y is finite.
75
Remark 1.7.9. If X is finite, then every ultrafilter on X is principal.
In the interesting case (X infinite), do non-principal ultrafilters exist? As-
suming AC (or “Zorn’s Lemma”), they do; AC is actually slightly stronger than
necessary to show this. In fact, there is an Ultrafilter Axiom which is inde-
pendent of ZF (and weaker than AC) saying that every filter is included in an
ultrafilter. Applying this to the Fréchet filter on X (Example 1.7.4), we find
non-principal ultrafilters. This is however highly non-constructive.
Lemma 1.7.10 (AC, or “Ultrafilter Axiom”). Every filter is included in an
ultrafilter.
(mi ) ∼ (ni ) if {i ∈ I : mi = ni } ∈ U
76
Similarly for functions:
∗
fM (m1i ) , . . . , (mki ) = (mk+1
i )
if
Mi
m1i , . . . , mki = mk+1
i∈I:f i ∈U
Checking that these are well-defined is very similar to showing the transitiv-
ity of the equivalence relation: if we have two equivalent sequences, then they
must agree on a set in U, which means that the functions or relations must agree
on a set containing a set in U, hence lying in U.
Theorem 1.7.16 (Łoś). Let ϕ be a wff. Let s∗ : V → M ∗ be an assignment,
say s∗ (x) = [(mi,x )i∈I ], and let si (x) = mi,x ; each si is an assignment V → Mi .
Then M∗ |= ϕ[s∗ ] iff {i ∈ I : Mi |= ϕ[si ]} ∈ U.
Proof . By induction on ϕ.
• If ϕ is atomic then it is clear from our definition of M∗ .
• Now suppose that ϕ is ¬ψ. If M∗ |= ¬ψ[s∗ ], then M∗ 6|= ψ[s∗ ]. By
induction, this means that {i ∈ I : Mi |= ψ[si ]} 6∈ U, and since U is an
ultrafilter (so it always contains a set or its complement), this means that
{i ∈ I : Mi 6|= ψ[si ]} ∈ U, i.e. that {i ∈ I : Mi |= ¬ψ[si ]} ∈ U. Hence
M∗ |= ϕ[s∗ ]. All of these steps are reversible, so the iff holds.
• Next assume that ϕ is ψ ∧ χ. Then:
M∗ |= (ψ ∧ χ)[s∗ ]
iff M∗ |= ψ[s∗ ] and M∗ |= χ[s∗ ]
iff {i ∈ I : Mi |= ψ[si ]}, {i ∈ I : Mi |= χ[si ]} ∈ U
iff {i ∈ I : Mi |= ψ[si ]} ∩ {i ∈ I : Mi |= χ[si ]} ∈ U
iff {i ∈ I : Mi |= (ψ ∧ χ)[si ]} ∈ U
77
1.7.3 Alternate Proof of Compactness*
Proof of compactness of first-order logic, Theorem 1.6.4. Fix a finitely
satisfiable first-order theory Σ. Let I be the collection of finite subsets of Σ.
For each i ∈ I, there is a structure Mi satisfying Σi by the definition of finite
satisfiability.
For each ϕ ∈ Σ, let Aϕ = {i ∈ I : Mi |= ϕ}. Then the family {Aϕ :
ϕ ∈ Σ} has the finite intersection property, since the intersection of a finite
family corresponds to Aϕ1 ∧···∧ϕn , which must be nonempty since Σ is finitely
satisfiable. By QLemma 1.7.13, there is an ultrafilter U extending {Aϕ : ϕ ∈ Σ}.
We let M∗ = Mi /U (i.e. M∗ is the ultraproduct of the Mi ’s with respect
to U).
Let ϕ ∈ Σ. We have (using Łoś’ Theorem and our definitions):
M∗ |= ϕ ⇔ {i ∈ I : Mi |= ϕ} ∈ U ⇔ Aϕ ∈ U,
1.8 Exercises
Exercise 1.1. State and prove a version of the Unique Readability Theorem
for terms of a first-order language.
Exercise 1.2. Let L be a purely relational first-order language (i.e., the signa-
ture consists only of relation symbols). State and prove the Unique Readability
Theorem for L-first-order wff’s.
Exercise 1.3. Let ϕ be the wff (∀v1 v1 = v2 ) → (∀v2 v2 = v1 ).
1. Is v3 substitutable for v1 in ϕ? If yes, give ϕ[v3 /v1 ].
2. Is v1 substitutable for v1 in ϕ? If yes, give ϕ[v1 /v1 ].
3. Is v2 substitutable for v1 in ϕ? If yes, give ϕ[v2 /v1 ].
78
4. Let t = f (v1 , v2 ). Find a suitable ϕ̃ given by Theorem 1.3.6, when one
wishes to replace v1 by t. Then give ϕ̃[t/v1 ].
Exercise 1.4. Let L be the empty language (pure equality).
1. Write an axiom saying that there at least n elements.
2. Show that this is not true for infinite M: find two infinite L-structures M
and N which satisfy the same sentences, but with different cardinalities.
Exercise 1.7. Let E be a binary relation symbol and consider the first-order
language L = {E}.
1. Write a (finite!) set of axioms T expressing that E is an equivalence
relation.
From now on we’ll implicitely be working with extensions of T (theories
extending T ). So it makes sense, for us, to speak about equivalence classes
(we say classes, for short).
2. Write an axiom saying that every class is reduced to a single element.
79
8. Write an axiom saying that there are exactly m classes with exactly n
elements.
9. Write a theory saying that every class is infinite.
10. Is there a theory satisfied exactly by (sets equipped with) an equivalence
relation with an infinite class?
No justification required.
Exercise 1.8. Let 0 be a constant symbol, s a unary function symbol and
consider the following theory:
∃x ∈ M ∀y ∈ M S(x) → S(y)
80
Exercise 1.12. Suppose the ∀ rules, the axiom v1 = v1 , for any function symbol
f the axiom
Sketch a proof (but no deductions, it would be too long!) of the fact that
for any families of terms (si ) and (ri ) and any formula ϕ in which they are
substituable for the vi ’s, ` (r1 = s1 ∧ · · · ∧ rn = sn ) → (ϕ[ri /vi ] ↔ ϕ[si /vi ]).
Just decompose the proof into steps, but do not give any details!
Exercise 1.13. Finish the proof of Proposition 1.4.2 by showing that = is
transitive, that is show ` ∀x ∀y ∀z (x = y ∧ y = z → x = z).
Exercise 1.14. Let ϕ be a wff, t a term, x a variable, and ϕ̃ given by the
renaming algorithm (Theorem 1.3.6) such that t is substitutable for x in ϕ̃.
Show (we haven’t done it yet) that {ϕ} ` ϕ̃ and {ϕ̃} ` ϕ.
Exercise 1.15. Formalize first-order logic using ∃, and consider ∀ as an abbre-
viation. Show that this is consistent syntactically and semantically.
Exercise 1.16. Prove the soundness theorem by induction with both quanti-
fiers.
Exercise 1.17 (Gödel translation, continued). We pursue in the vein of Exer-
cise 0.19. Define:
• ϕ∗0 = ϕ0 for ϕ0 an atomic formula;
• (∀x ϕ)∗ = ∀x ϕ∗
Prove the same conclusions as in Exercise 0.19.
Exercise 1.18. For n ∈ N, let ϕn be the formula ∃v1 . . . ∃vn ∧i6=j vi 6= vj . Let
Σ = {ϕn }. Show that Σ 6` ϕn+1 .
Exercise 1.19. The language is the language of groups.
1. Give a finite theory whose models are all groups.
2. Give a theory whose models are exactly all infinite groups.
3. Find a finite theory whose models are infinite groups (requires some cul-
ture).
4. Show that there is no theory whose models are exactly all finite groups.
Exercise 1.20. The language has a binary relation. Prove that there is no
theory whose models are all connected graphs.
81
Exercise 1.21 (definable sets). Let M be a first-order structure in a first-order
language L.
1. A subset X ⊆ M is definable if there is a formula ϕ with FreeVar ϕ = {v0 }
such that for all assignment s: s(v0 ) ∈ X iff M |= ϕ[s].
(Intuitive meaning: m is in X iff ϕ is true of m.)
Show that {0} is definable in (N, <). Show that {1} is definable in (N, <).
Show that the set of prime numbers is definable in (N, 0, 1, +, ·).
2. A subset X ⊆ M is definable with parameters µ = (µ1 , . . . , µk ) ∈ M k
if there is a formula ϕ with FreeVar ϕ = {v0 , . . . , vk } such that for all
assignment of the variables s mapping vi to mi (i = 1 . . . k): s(v0 ) ∈ X iff
M |= ϕ[s].
(Intuitive meaning: m is in X iff ϕ(m, µ1 , . . . , µk ) is true.)
√
Show that { π} is definable with parameters π in (R, 0, 1, +, −, ·).
Show that R> 0 is definable in (R, 0, 1, +, −, ·).
3. A subset X ⊆ M n is definable with parameters µ ∈ M k if there is a for-
mula ϕ with FreeVar ϕ = {x1 , . . . , xn , y1 , . . . , yk } such that for all assign-
ment of the variables s mapping yi to µi (i = 1 . . . k): (s(v1 ), . . . , s(vn )) ∈
X iff M |= ϕ[s].
Show that the function ÷ is definable in (R, 0, 1, ·). (i.e., show that its
graph is definable.)
4. Bonus questions: R is not definable in (C, 0, 1, +, ·). Z is not definable
in (R, 0, 1, +, ·). All definable subsets of C in (C, 0, 1, +, ·) are finite or
cofinite.
Exercise 1.22.
1. Show that {k ∈ Z : k is odd } is definable in (Z, +, ·).
2. Show that each element of Q is definable in (R, 0, 1, +, −, ·).
3. Show that {i, −i} is definable in (C, 0, 1, +, −, ·). Show that {i} is not.
4. Show that R is not definable in (C, 0, 1, +, −, ·).
Exercise 1.23 (non-standard models of (N, s)). Let L = {0, s} where 0 is a
constant symbol and s a unary function; let c be another constant symbol and
L0 = L ∪ {c}. Let T consist of all sentences true in (N, 0, s) where s is the
successor function.
1. Let T 0 = T ∪ {c 6= s| ◦ ·{z
· · ◦ s} : n ∈ N}. Show that T 0 is consistent.
n times
2. Let N0 be a model of T 0 . Construct a canonical injective function N ,→ N0 .
Is it surjective?
82
3. Let X ⊆ N0 be a definable, non-empty subset (definable by a formula with
no parameters). Show that X has a least element.
Exercise 1.24. Let M be a finite first-order structure. Let T be the collection
of all L-sentences true in M. Show that all models of T have same cardinal as
M. (They are actually all isomorphic to M, whatever that means.) Show that
this fails if M is infinite.
Exercise 1.25.
1. Let T be a complete theory. Suppose that T has an infinite model. Show
that all models of T are infinite.
83
Exercise 1.30 (universal closures). Let ϕ be a formula in a first-order language
L, and let FreeVar ϕ = {x1 , . . . , xn }. We define the universal closure of ϕ to be
Let σ be the sentence : f (a) < b. Prove that T 6` σ and that T 6` ¬σ.
84
Chapter 2
Second-Order Logic
(Presto)
In this short chapter we briefly describe second-order logic, in which one may
quantify on collections of elements. This requires new symbols ∀2 and ∃2 . The
logic is amzazingly expressive, too expressive to have interesting results.
In this chapter:
• Explain what second-order means.
85
One should define terms and wff’s, but we do not wish to repeat these
discussions, as we shall not try to describe a proof theory. There are two reasons.
• First, we would have to determine suitable axioms for the relationship
between predicates and elements, as we determined reasonable axioms
for equality in the case of first-order logic. This amounts to agreeing on
axioms for membership, that is agreeing on a theory of sets, which is not
the spirit of this course.
• Second, and worse, this appears a little pointless to us, as compactness
will fail anyhow: hence so would completeness.
We therefore entirely forget about proof theory.
injects(X 1 , Y 1 ) : ∃2 Γ2 injection(Γ2 , X 1 , Y 1 )
86
One should compare this to first-order logic, in which “infinite” is not a single
formula, but an infinite theory not finitely axiomatizable (Example 1.6.7). Exercise 2.1
• s is an injective function
• every element not 0 is in the image of s
• there are no cycles, i.e. for each n the axiom:
∀x s| ◦ ·{z
· · ◦ s}(x)¬x
n
• Every subset which contains 0 and is closed under s is the entire base set:
87
Proof . Let M |= PA. We find a bijection between N and M which maps 0 to
0M and which is compatible with succession. For the sake of clarity, we write
n + 1 for sN (n) (succession in N happens to coincide with adding 1, though
addition is not part of the language).
Let f (0) = 0M . Inductively, we now define f (n + 1) = sM (f (n)). This does
define a function N → M , which is by definition compatible with s. Clearly f
is injective (induction in N; s has no cycles in M).
It remains to show that f is surjective. Consider im f ⊆ M . This subset
contains 0M and is closed under sM ; as M satisfies the axioms of PA, one has
im f = M . So f is surjective; it is a successor-preserving bijection.
Lecture 19 (ME2)
3. of cardinality ℵ1
4. of cardinality ℵn
5. of cardinality continuum (that is 2ℵ0 , like R).
Hint: for the last one, use the following algebraic fact. There is up to isomor-
phism a unique ordered field which is complete.
88
Chapter 1’
89
have been interested in proof theory; model theory is on the semantic side, so we
shall make notations more convenient, and disregard the risk of an ambiguity.
1’.0.2 Quantifications
We rehabilitate poor ∃ and make it a quantifier again.
Definition 1’.0.5 (quantifier-free formula). A formula is quantifier-free if it has
no quantifiers.
Definition 1’.0.6 (universal formula). A formula is universal if it is of the form
∀v1 . . . ∀vn ϕ0 , where ϕ0 is quantifier-free.
Definition 1’.0.7 (existential formula). A formula is existential if it is of the
form ∃v1 . . . ∃vn ϕ0 , where ϕ0 is quantifier-free.
90
Be very careful that ∀v1 ∃v2 . . . is neither universal nor existential.
91
1’.1 Elementary Equivalence; Inclusion and El-
ementary Inclusion
Remark 1’.1.1. “Elementary” is the standard adjective used in model theory;
one must be careful with the definitions (elementary equivalence, Definition
1’.1.5, and elementary inclusion, Definition 1’.1.13), as the role of parameters
is not necessarily obvious. Terminology tends to be confusing; “elementary”
captures a way of thought, not a notion.
In particular, Th M = Th(M, ∅). One may also notice that Th0 (M, A) =
Th(M, A) ∩ {ϕ ∈ LA : qrk(ϕ) = 0}.
Remark 1’.1.4. If M is an L-structure and A ⊆ M , then Th(M, A) is a
complete LA -theory.
Lemma 1’.1.6. A theory T is complete iff for all models M, N of T , one has
M ≡ N.
92
Proof . Recall that we have identified T with the set of its consequences, and
that a theory is complete iff it is maximal as a consistent set of sentences.
Suppose T complete. If M |= T , then T ⊆ Th M. As T is complete, one
has T = Th M; since T = Th N similarly, it follows M ≡ N .
Suppose that any two models of T are elementarily equivalent. Let ϕ be a
sentence such that neither ϕ not ¬ϕ is in T . Then there are models M |= T ∪{ϕ}
and N |= T ∪ {¬ϕ}. In particular, M ≡ N , and this is a contradiction. So
either ϕ or ¬ϕ is in T .
So the running question: is this theory complete? Has a partial answer: can
one show that any two of its models are elementarily equivalent?
1’.1.2 Inclusion
We now turn our attention to substructures; the naive notion (Definition 1’.1.7)
is not very useful, the interesting form will be elementary inclusion (Definition
1’.1.13). As noted in Remark 1’.1.1, terminology might be a bit confusing, as “el-
ementary” will not exactly mean the same thing as in “elementary equivalence”
(Definition 1’.1.5).
But let us begin with the naive, not-so-useful, notion of inclusion.
Definition 1’.1.7 (substructure). Let M, N be L-structures. Then M is a
substructure of N , written M ⊆ N , if the following hold:
• M ⊆ N;
93
• for each function symbol f , M is closed under f N .
The structure M is then uniquely determined.
When studying Löwenheim-Skolem theorems we shall say a little more about
generating substructures (§1’.3.2).
The following is a very healthy exercise.
Remark 1’.1.10. Let I be a non-empty set and (Mi )i∈I be an increasing chain
of L-substructures, that is i ≤ j ⇒ Mi ⊆ Mj . Then ∪i∈I Mi naturally bears
an L-structure M∗ such that for all i ∈ I, Mi ⊆ M∗ .
(ii). Induction again. If there are no quantifiers, then this is known. Suppose
that ϕ is ∃x ψ(x, m); ψ is existential again. If M |= ∃x ψ(x, m), then
there is µ ∈ M such that M |= ψ(µ, m). By induction, N |= ψ(µ, m); so
N |= ϕ(m).
94
1’.1.3 Elementary Inclusion
We now introduce a better-behaved notion of inclusion (and substructure). The
intuitive idea is that both structures should agree on as many properties as
possible, allowing parameters. Of course in order for the question to make
sense, only elements of the small structure can be taken into account.
Definition 1’.1.13 (elementary substructure). Let M ⊆ N be an inclusion
of L-structures. M is an elementary substructure of N , written M N , if
Th(M, M ) = Th(N , M ).
Example 1’.1.14.
• (N, <) 6 (Z, <), as N |= ∀x x ≥ 0 but Z |= ∃x x < 0. (This formula uses
0 as a parameter.)
• (Q, <) (R, <) (not necessarily obvious yet).
• Let M ⊆ N be infinite sets, regarded as pure sets. Then M N .
• We have observed in Counter-example 1’.1.12 that none of the inclusions
of rings Z ⊆ Q ⊆ R ⊆ C is elementary.
• As rings, Q C (not necessarily obvious).
• When we constructed non-standard reals in §1.6.3, we considered a the-
ory extending Th(R, R), in order to force the inclusion R ⊆ R∗ to be
elementary.
In other words, M N iff M and N have the same LM -theory, iff N |=
Th(M, M ). This is also equivalent to: for any LM formula ϕ(m) with parame-
ters in M , M |= ϕ(m) iff N |= ϕ(n).
In particular M N implies M ≡ N , but the converse need not hold.
Counter-example 1’.1.15. Consider N = (Z, <) and M = (2Z, <) ⊆ N . As
they are clearly isomorphic (Definition 1’.2.6 below), it is clear that they are
essentially the same object, and M ≡ N (this will be more convincing later).
Yet M |= ∀x (x ≤ 0 ∨ x ≥ 2), and N |= ∃x (0 < x < 2). Hence M and N can
be distinguished by the tuple (0, 2).
Lemma 1’.1.16. Let I be a non-empty set and (Mi )i∈I be an increasing el-
ementary chain of L-substructures, that is i ≤ j ⇒ Mi Mj . Let M∗ =
∪i∈I Mi . Then for all i ∈ I, Mi M.
Proof . A very healthy exercise. Exercise 1’.1
Ex 1’.4, 1’.5
The following gives a criterion for elementarity of inclusions; it should be
used mostly in order to go down, that is, when one already has perfectly under-
stood the notion of satisfaction in N , and tries to see if M ⊆ N is an elementary
substructure. (This will typically be the case in §1’.3.2). The statement is more
subtle than it first seems.
95
Theorem 1’.1.17 (Tarksi’s test). Let M ⊆ N be an inclusion of L-structures.
Then M N iff for all formula ϕ(m) with parameters in M ,
96
1’.2.1 Morphisms
Notation 1’.2.1. If m = (m1 , . . . , mn ) is a tuple and σ : M → N is a function,
we write σ(m) for the tuple (σ(m1 ), . . . , σ(mn )).
Definition 1’.2.2 (L-morphism). Let M, N be L-structures and σ : M → N
be a function. σ is an L-morphism if:
• σ(cM ) = cN for any constant symbol c;
• m ∈ RM iff σ(m) ∈ RN for any relation symbol R;
• σ(f M (m)) = f N (σ(m)) for any function symbol f .
In short, an L-morphism is a function which is compatible with the interpreta-
tions of L in M and N .
In particular, as equality is always among the relations, one sees that an L-
morphism is always injective. This is the reason why one also says L-embedding.
Remark 1’.2.3. The composition of two L-morphisms is an L-morphism; IdM
is always an L-morphism.
Example 1’.2.4.
• Let M and N be orderings. Then σ : M → N is an {<}-morphism iff it
is an order-preserving (injective) function.
• An injective group homomorphism is an Lgrp -morphism.
• Any ring morphism between fields is an Lrng -morphism.
Remark 1’.2.5.
• If M ⊆ N , then the inclusion map M → N is an L-morphism.
• More precisely, let M, N be two structures with M ⊆ N (the under-
lying sets). Then M ⊆ N as L-structures iff the inclusion map is an
L-morphism.
• Let M, N be L-structures. Then there is an L-embedding M → N iff N
can be turned into a model of Th0 (M, M ) (theory of M with no quanti-
fiers, parameters in M ).
One may rush to the definition of an isomorphism.
Definition 1’.2.6 (L-isomorphism).
• A L-morphism is an L-isomorphism if it is surjective.
• M and N are L-isomorphic, written M ' N (L being implicit) If there
is an L-isomorphism between them.
(Recall that an L-morphism is always injective!)
97
Remark 1’.2.7. The composition of two L-isomorphisms, the reciprocal map-
ping of an L-isomorphism, are L-isomorphisms.
Example 1’.2.8. (Z, <) and (2Z, <) are clearly isomorphic.
Remark 1’.2.5 however suggested that we did not take the good definition in
the first place: morphisms reduce to inclusions, which in general have but little
interest. So we look for something more reminiscent of the notion of elementary
inclusion.
98
Corollary 1’.2.13. If M ' N , then M ≡ N .
Proof . Th(M) is the set of properties of the empty tuple ∅ in M. As there is
an isomorphism σ, one has that for any sentence, M |= ϕ iff N |= ϕ.
1’.2.3 Categoricity
Definition 1’.2.14 (categoricity). Let T be a first-order theory with infinite
models and κ an infinite cardinal. T is κ-categorical if it has a model of cardi-
nality κ and any two models of T of cardinal κ are isomorphic.
99
1’.3 Löwenheim-Skolem Theorems
1’.3.1 The Substructure Generated by a Set
Notation 1’.3.1. If L is finite, we write Card L = ℵ0 anyway.
Notice that there is no risk of confusion if Card L ≥ ℵ1 . You may think that
in any case, Card L actually denotes Card WFFL .
Let us start with an L-structure M and a subset A of M . One can easily
find the smallest substructure of M containing A: it suffices to close A (union
the constants) under the functions of the language; the resulting substructure
has cardinality at most Card A + Card L.
(The idea is clear. One wishes to add the images of elements of A0 through
functions of L. Unfortunately f (A0 ) does not necessarily make sense, as the
arity of the function may vary. The union indexed over N deals with this minor,
strictly notational, issue.)
From A1 we build A2 in a similar fashion, and keep going. Eventually the
set hAi = ∪n∈N contains A, all (interpretations of) constants, is closed under
(the interpretations of) all functions in the language. We interpret relation
symbols on hAi by the induced interpretation (i.e., as a subset of M ), getting
an L-structure which is a substructure of M.
Notice that A0 has cardinality at most Card A + Card L; at each stage, we
add at most Card L elements; there are ℵ0 ≤ Card L stages, and this proves
CardhAi ≤ Card A + Card L.
Example 1’.3.4. Consider the case of a subset S of a group G again. Then
S0 = S ∪ {1}, S1 = S0 ∪ S0−1 ∪ S0 · S0 , etc. One does find hSi = ∪n∈N Sn .
Definition 1’.3.5 (substructure generated by a set). Let M be an L-structure
and A ⊆ M be any set. The substructure generated by A in M is the smallest
substructure of M containing A. It coincides with hAi from Lemma 1’.3.3.
100
One can prove that an arbitrary intersection of substructures of M is a
substructure of M again; we have just been generalizing to L a construction
which is well-known for groups. It follows that hAi is the intersection of all
substructures of M containing A. Notice in any case that hAi is canonical, and
that we could construct it explicitely.
All this is very nice, but in general hAi has no reason to be an elementary
substructure of M (Definition 1’.1.13), and this is precisely what we would be
interested in. So Lemma 1’.3.3 and Definition 1’.3.5 are not that useful.
The meaning is clear: our new axioms say that witnesses are explicitely given
by our new function symbols.
Lemma 1’.3.8. Let M be a model of T . Then there is an interpretation of the
Skolem functions making M a model of TSk .
Proof . We must give a meaning to the Skolem function symbols fϕ . Let ϕ(x, y)
be an L-formula, and m ∈ M have same length as y. If M |= ∃x ϕ(x, m), then
there is µ ∈ M such that M |= ϕ(µ, m); we choose such a µ and set fϕM (m) = µ.
If there is no suitable µ, we map m to any element in M .
With heavy use of the axiom of choice (we perform infinitely many choices),
we thus define functions fϕM , which are the interpretations of the Skolem func-
tion symbols. By construction, M |= ∀y ∃x ϕ(x, y) → ϕ(fϕ (y), y) .
Let us return to our problem. If we now close a subset A ⊆ M under all
functions including the Skolem functions, we are in good position to find an
elementary inclusion, as all necessary witnesses have been plugged in.
101
Theorem 1’.3.9 (descending Löwenheim-Skolem theorem). Let M be an infi-
nite L-structure and A ⊆ M . Let κ be a cardinal with Card A + Card L ≤ κ ≤
Card M . Then there is M0 M such that A ⊆ M0 and Card M0 = κ.
Proof . We may assume that Card A = κ. We turn M into an LSk -structure
which satisfies TSk . In the language LSk , we now consider the substructure of
M generated by A (Lemma 1’.3.3). Forgetting the interpretation of the Skolem
functions, this is a fortiori an L-structure M0 . Moreover Card M0 = κ.
It remains to prove that the inclusion is elementary. We use Tarski’s Test
(Theorem 1’.1.17). Let ϕ(m0 ) be a formula with parameters in M0 ; assume
M |= ∃x ϕ(x, m0 ). As M |= TSk , one has M |= ϕ(fϕM (m0 ), m0 ). But by
construction, the element fϕM (m0 ) lies in M0 . This is what we need in order to
apply Tarski’s criterion.
Remark 1’.3.10.
• M0 M as L-structures, not necessarily as LSk -structures. This is
because when we expand the language to LSk , we create new formulas. If
we did want to ensure M0 M as LSk -structures, we would need to add
new Skolem functions for the new formulas, getting LSk,Sk , etc.
However an inclusion of L-structures is what we want; the expanded lan-
guage LSk is a technical device we can forget about.
• The construction is highly non-canonical (it relies on the axiom of choice).
In general, there is no smallest elementary substructure containing A:
everything depends on the way we interpreted the Skolem functions, which
is arbitrary. One should not use the notation hAi for “Skolem hulls”.
As a final interesting word, we prove inconsistency of set theory.
Corollary 1’.3.11 (Skolem’s paradox). (If set theory is satisfiable) there is a
countable model of set theory.
Proof . Write down ZFC as a first-order theory in the language {∈}. Start with
a model and apply Theorem 1’.3.9.
However in this model U0 of set theory we may construct N and P (N),
which is notoriously uncountable, and yet a subset of the countable underlying
universe. . .
[Catch: “countable” does not bear the same meaning in the first model, say
U, and the countable one U0 . All that we have proved is that U0 is so small that
it does not “see” (i.e., perceive as a set) a bijection between P (N) and N.]
102
Theorem 1’.3.12 (ascending Löwenheim-Skolem theorem). Let M be an infi-
nite L-structure and κ ≥ Card M be a cardinal. Then there is M∗ M such
that Card M ∗ ≥ κ.
Proof . Let C 0 = {ci : i < κ} be new constants, and consider the language
L0 = LM ∪ C 0 . Form the L0 -theory T 0 = Th(M, M ) ∪ {ci 6= cj : i 6= j}. This
theory is consistent: a finite fragment mentions a finite number of properties
ϕ(m), which are trivially satisfiable in M, and a finite number of formulas
ci 6= cj , meaning that there are at least n distinct elements: this is satisfiable
in M again.
Hence any finite fragment of T 0 is satisfiable in M; by compactness, there is a
a model M∗ of T 0 (notice that M∗ is not likely to be M). As M∗ |= Th(M, M ),
one has M M∗ ; on the other hand, the constants of C 0 say that M∗ has at
least κ elements.
Combining Theorems 1’.3.9 and 1’.3.12, one derives a sharper version.
Theorem 1’.3.13 (Löwenheim-Skolem theorem). Let M be an L-structure and
A ⊆ M . Let κ be a cardinal with Card A + Card L ≤ κ. Then there is M0 such
that A ⊆ M 0 , Card M 0 = κ, and M0 M or M0 M.
Proof . Go up (Theorem 1’.3.12), then down (Theorem 1’.3.9).
Corollary 1’.3.14 (Łoś-Vaught test). Let T be a first-order theory in a language
L. If there is κ ≥ Card L such that T is κ-categorical, then T is complete.
Proof . Let M, N |= T . We aim at showing M ≡ N .
By the Löwenheim-Skolem theorem (Theorem 1’.3.13), there are M0 and N 0
of cardinal κ such that M M0 or M M0 , and similarly for N and N 0 . But
in any case M ≡ M0 and N ≡ N 0 . In particular, M0 , N 0 |= T .
As M0 and N 0 are models of T of cardinal κ, they are isomorphic by cate-
goricity; it follows M0 ≡ N 0 . As a conclusion, M ≡ M0 ≡ N 0 ≡ N .
Example 1’.3.15. We shall prove later (Theorem 1’.4.4) that DLO is ℵ0 -
categorical. In particular it will be complete!
103
methods which enable to inductively extend partial functions. One should bear
in mind the image of a bootstrap.
Disclaimer: back-and-forth methods are easier to work out in the case of
relational languages. With unary functions they are still manageable; if there
is a binary function around, it is not very likely that one will understand any
kind of back-and-forth.
104
• at odd stages, we shall make sure that σ is defined everywhere on M ;
• at even stages, we shall make sure that σ reaches every element of N .
We thus want to find an increasing sequence of partial maps which preserve
the ordering. Let us start by defining σ0 (m0 ) = n0 . So far, so good, as all we
can say is m0 = m0 , which certainly holds of n0 .
• (stage 2k+1) Suppose that σ2k is a partial map with finite support, and
such that a = dom σ2k ∈ M and b = im σ2k ∈ N satisfy the same relations.
(You may think that σ2k “preserves the configuration of the tuple”, that
it is a partial L-morphism in the sense of Definition 1’.2.2.) We want to
extend σ2k to an order-preserving map σ2k+1 with mk+1 ∈ dom σ2k+1 .
mk+1 is in a certain configuration with respect to the finite tuple a. If
mk+1 lies above any element of a, then as there are no endpoints in N ,
there is certainly β ∈ N which lies above any element of b. If on the other
hand, mk+1 lies between two elements of the tuple a, then as N is dense,
the same pattern can be reproduced in N . In any case there is β ∈ N such
that the extended tuples a, mk+1 and b, β still satisfy the same relations.
We let σ2k+1 extend σ2k by σ2k+1 (mk+1 ) = β.
• (stage 2k + 2) Suppose that σ2k+1 has been constructed as a finite partial
map which preserves relation between a0 = dom σ2k+1 and b0 = im σ2k+1 .
We aim at extending σ2k+1 in such a way that nk+1 will be in the image.
And we argue exactly like above, getting in any case an element α ∈ M
such that α behaves with respect to a0 exactly the way nk+1 does for b0 .
We let σ2k+2 (α) = nk+1 .
At the limit, we let σ = ∪k∈N σk . This is still an order-preserving bijection.
Because of our construction at odd stages, every element of M will appear in the
domain, so dom σ = M . Because of our construction at even stages, im σ = N .
It follows that σ : M → N is an isomorphism.
Bearing in mind Definition 1’.2.14, one sees that Theorem 1’.4.4 says that
DLO is ℵ0 -categorical. As a consequence, one can deduce from Corollary 1’.3.14
that DLO is complete.
105
• Above all, the proof relies on this bootstrap construction, arguing that
any finite configuration on one hand can be reflected on the other hand.
This is worth a general definition (Definition 1’.4.10 hereafter).
1’.4.2 ∞-isomorphisms
Definition 1’.4.6 (local isomorphism, 0-isomorphism).
• A local isomorphism, or 0-isomorphism, σ : M → N is an injective partial
function M → N with finite domain which preserves the constants, func-
tions, and relations. Hence σ is a local isomorphism iff dom σ and im σ
satisfy the same atomic formulas (in M, N respectively).
• Two tuples a ∈ M and b ∈ N are locally, or 0-isomorphic, written a '0 b,
if there is a local isomorphism σ mapping a to b.
• Two structures M and N are locally, or 0-isomorphic, written M '0 N ,
if the empty function ∅ 7→ ∅ is a 0-isomorphism.
“Local” is of course, opposed to “global”: we consider only functions with
finite support. So a local isomorphism is a small fragment of a morphism. When
one wants to insist on the poor quality of σ, one calls it a 0-isomorphism: it
preserves only formulas of quantification rank 0. This is very far from being
elementary!
Remark 1’.4.7. A beginner’s mistake is to think that the empty tuple of M
is always 0-isomorphic to the empty tuple from N . This is not the case, as
there might be constants in the language. For instance, the ring Z/2Z satifies
1 + 1 = 0, but the ring Z/3Z doesn’t. So even the empty function is not
compatible with the interpretation!
A little thinking will reveal that M '0 N iff Th0 (M) = Th0 (N ) (quantifier-
free theories, no parameters: this boils down to atomic formulas using terms
built from constants).
Example 1’.4.8.
`
• Consider the orderings (Z, <) and`(Z < Z, <); we denote by 0 the 0 of
Z; there are two copies of 0 in Z < Z, which we denote respectively by
01 and 02 . Clearly ∅ '0 ∅. Clearly too, 0 '0 01 . Also, (0, 2) '0 (01 , 02 )
(as 0 < 2 and 01 < 02 ).
`
• Consider the successor function and the structures (Z, s) and (Z Z, s).
Clearly ∅ '0 ∅ and 0 '0 01 . But it is no longer the case that (0, 2) '0
(01 , 02 ). Indeed, M |= 2 = s(s(0)), whereas N |= 02 6= s(s(01 )).
Example 1’.4.9. A moment’s thought and one realizes that in the language of
orderings, local isomorphisms are exactly order-preserving bijections with finite
support.
We can now give the key definition. Bear in mind the image of a bootstrap.
106
Definition 1’.4.10 (∞-isomorphism, Karp family). Let M and N be two L-
structures. M and N are ∞-isomorphic, written M '∞ N , if there is a non-
empty family F of local isomorphisms M → N such that:
forth: for all σ ∈ F and α ∈ M , there is τ ∈ F extending σ such that α ∈ dom τ
back: for all σ ∈ F and β ∈ M , there is τ ∈ F extending σ such that β ∈ im τ
F is called a Karp family, and its elements are called ∞-isomorphisms.
Remark 1’.4.11.
• If there is a Karp family, there is a largest one (take the union of all).
• M ' N implies M '∞ N (take all restrictions of a global isomorphism),
but the converse need not hold in general: perhaps M and N don’t even
have the same cardinality (Counter-example 1’.4.13 below).
Proposition 1’.4.12. Any two DLO’s are ∞-isomorphic. More specifically, the
family of local isomorphisms is a Karp family.
Proof . Local isomorphisms between orderings are just order-preserving, finite
support bijections (which do exist: the family is not empty!). The back-and-
forth property is the core of the proof of Theorem 1’.4.4.
We emphasize that ∞-isomorphic is much weaker than isomorphic.
Counter-example 1’.4.13. Consider the orderings (Q, <) and (R, <). They
are ∞-isomorphic, but of course Q 6' R.
There is however a partial correction.
Proposition 1’.4.14. If M '∞ N and both are countable, then M ' N .
Proof . This is an abstract version of the bootstrap construction used in The-
orem 1’.4.4.
Propositions 1’.4.12 and 1’.4.14 together yield the proof of Theorem 1’.4.4.
Counter-example 1’.4.15. The notions of ∞-isomorphism and of a Karp
family (Definition 1’.4.10) are about arbitrarily large finite configurations; one
should therefore not expect to generalize Proposition 1’.4.14 past ℵ0 , even if M
and N have the same cardinality. Here
` is a counter-example.
Consider
` an ordered sum M = ℵ1 Q (copies of Q copied after each other),
and N = ℵ1 (Q ∪ {∞}) (Q followed by a point). Both are dense linear order-
ings without endpoints, of cardinality ℵ1 . They are ∞-isomorphic (Proposition
1’.4.12), and yet they are not isomorphic. The latter is non-trivial.
Back-and forth arguments are especially powerful in ω-saturated models
(defined below, Definition 1’.5.12). For instance, it is not true that any two
algebraically closed fields are ∞-isomorphic, but any two ω-saturated models
(defined below, §1’.5.2) are. These are algebraically closed fields of infinite tran-
scendence degree; you may already try to establish a back-and-forth between
them. This will be done in §1’.5.3.
107
1’.4.3 Finitary Back-and-Forth*
We build on local isomorphisms to find weaker versions of back-and-forth.
Definition 1’.4.16 (n + 1-isomorphism).
• A local isomorphism σ : a → b is an (n + 1)-isomorphism if:
Here begin the funny games: given two structures, how isomorphic are they?
Remark 1’.4.18. Back-and-forth may be viewed as a game opposing two play-
ers, having two different boards M and N . A round consists of the following
two steps:
108
Player 2 will identify Z with Z1 : whenever player 1 makes a move in Z,
player 2 does the same in Z1 ; whenever player 1 makes a move in Z1 ,
player 2 does the same in Z. Now, when player 1 plays in Z2 , player 1
picks a similar configuration in very big elements of Z, of size around 2k .
Player 1 will win, but it will take him at least k rounds!
109
Proof . By induction. First notice that by definition (Definition 1’.4.6), if a '0
b, then a and b satisfy the same relations, hence the same atomic formulas; the
claim is proved.
Suppose that the property is known for n; we show it for n + 1. So let
a 'n+1 b be two n + 1-isomorphic tuples. We show tpn+1 (a) = tpn+1 (b). Let ϕ
be a formula of tpn+1 (a). We show that N |= ϕ(b).
A quick induction on the structure of ϕ reveals that the only interesting case
is that of ϕ(a) being ∃x ψ(a, x), for a formula ψ now of quantification rank ≤ n.
We have assumed M |= ϕ(a): so there exists α ∈ M such that M |= ψ(a, α).
In particular, ψ ∈ tpn (a, α). As a and b are n + 1-isomorphic, there is by
definition (Definition 1’.4.16) β ∈ N such that a, α 'n b, β. By induction,
ψ ∈ tpn (b, β); that is, N |= ψ(b, β). It follows that N |= ϕ(b): we are done.
Remark 1’.4.25. We insist that if a and b are locally isomorphic, and that
there is a Karp family of extensions of a '0 b (one may then write a '∞ b), a
and b satisfy the same formulas with quantifiers.
Proof of Proposition 1’.4.22. Suppose M 'ω N ; we show M ≡ N . Our
assumption means that for each n, ∅M 'n ∅N . By Lemma 1’.4.24, tpM n (∅) =
tpN
n (∅). But tp M
n (∅) = Th n (M) (the theory of M restricted to formulas of
qrk ≤ n). As this is true for any n, one gets Th(M) = Th(N ); so M ≡ N .
Proposition 1’.4.22 does not have a converse, because when n > 0, Lemma
1’.4.24 doesn’t.
`
Counter-example 1’.4.26. Consider the case of (Z, s) and (Z Z, s) again.
Then 0 and 01 clearly satisfy the same formulas of qrk ≤ 1, but as observed,
they are not 1-isomorphic.
Remark 1’.4.27. The choice of Counter-example 1’.4.26 is typical. Actually, in
a language consisting of finitely many relations, the converse to Lemma 1’.4.24
does hold, and in particular M 'ω N iff M ≡ N !
We sum up the isomorphism strength as follows (n < m)
Counter-example 1’.4.28.
• (R, <) '∞ (Q, <) but they are not isomorphic (Counter-example 1’.4.13)
`
• (Z, <) 'ω (Z < Z, <) but they are not ∞-isomorphic (Remark 1’.4.21).
`
• (Z, s) ≡ (Z Z, s) but they are not ω-isomorphic (Example 1’.4.19).
`
• (Z, s) '1 (Z Z, s) but they are not 2-isomorphic (Example 1’.4.19).
110
• If M '∞ N are countable, then M ' N (Proposition 1’.4.14).
• There always is an ordinal such that M 'α N implies M '∞ N .
• In a finite, purely relational language, M ≡ N does imply N 'ω N .
111
Proof . One implication is trivial. For the other one, we may clearly assume
that E is closed under Boolean operations. (In particular, E contains a con-
tradiction.) Let ϕ(x) be an L-formula. We shall prove that ϕ(x) is equivalent
(modulo T ) to a formula of E.
Step 1. There is a finite subset Ψ ⊆ E such that the theory T ∪ {ψ(c) ↔ ψ(d) :
ψ ∈ Ψ} ∪ {ϕ(c) ∧ ¬ϕ(d)} is inconsistent.
Verification: Let c, d be tuples of new constant symbols (of the same length as
x) and L0 = L ∪ {c, d}. Let
112
assumption, ∅ satisfies the same formulas in M and N , that is M ≡ N . Yet
M |= ∀x (ϕ(x) ↔ ψM (x)), a contradiction. So T 0 is not consistent.
By compactness, there is a finite fragment of ΘM , which we can take to be
a single sentence θM , such that T ∪ {θM } |= ∀x (ϕ(x) ↔ ψM (x)). ♦
Step 4. Finitely many θM ’s suffice for all models of T .
Verification: Let T 0 = T ∪ {¬θM : M |= T } (this makes sense: the collection
of models is a proper class, but the collection of θM ’s is a set alright, as a
subcollection of the set of L-sentences). If T 0 were consistent, we would have a
model N |= T in which none of the θM ’s holds, which is impossible!
By compactness, T 0 is finitely inconsistent, and this means that there is a
finite set Θ of θM ’s such that T |= ∨Θ θM . ♦
Step 5. There is ψ(x) ∈ E such that T |= ∀x ϕ(x) ↔ ψ(x).
Remark 1’.5.5. This proof is a little long and clumsy. We have just been
covering a normal, compact space; a topological argument is possible, short,
and elegant, but it requires introducing spaces of types, which is a little beyond
the scope of a first introduction.
Exercise 1’.9
Corollary 1’.5.6. A theory T eliminates quantifiers if and only if, for all models
M and N of T and all tuples m and n extracted from M and N respectively,
if m and n satisfy the same atomic formulas, then they satisfy the
same formulas.
The question is now: how do we prove that two tuples do satisfy the same
formulas? In practice, this often results from a back-and-forth construction.
Corollary 1’.5.7. Let T be a first-order theory such that for any two models
M, N |= T , local isomorphisms between M and N form a Karp family. Then
T eliminates quantifiers.
113
Proof . We use Corollary 1’.5.6. Let m, n be drawn from M, N respectively
and satisfy the same quantifier-free formulas; we show that they satisfy the
same formulas. As m and n satisfy the same quantifier-free formulas, they are
locally isomorphic. By assumption, they are ∞-isomorphic. So m '∞ n, and
it follows tpM (m) = tpN (n) (see Remark 1’.4.25). Hence m and n satisfy the
same formulas; T eliminates quantifiers.
114
Definition 1’.5.9 (type). Let L be a first-order language, M an L-structure,
A ⊆ M a set of parameters. An n-type over A is a collection p(x) of LA -formulas
ϕ(x, a) (a ∈ A) which is consistent with Th(M, A).
Treating the variables as constants, one sees that an n-type is merely a
theory in LA (x) extending Th(M, A).
This generalizes the case of the type of a tuple (Definition 1’.4.23), which
is called realized. But any type is the type of a tuple which “hasn’t appeared
yet”: every type is realized somewhere.
Remark 1’.5.10. If p(x) is a type over A ⊆ M , then there is N M and
n ∈ N such that n realizes p. (Of course N will tend to be a proper extension
of M.)
Types are amazingly important and interesting (they come with nice topo-
logical methods), and students should refrain the teacher from going to deep
into this matter.
Example 1’.5.11.
• If A ⊆ M and b ∈ M , then tp(b/A) = {ϕ(x, a) : ϕ ∈ LA : M |= ϕ(b, a)} is
an n-type over A.
• Consider (N, <). Then {x > n : n ∈ N} is a 1-type over N. Notice that it
is not realized in N, but in some elementary extension.
• When we constructed the non-standard reals in §1.6.3, we simply con-
structed a 1-type over R, namely p(x) = {0 < x < r : r ∈ R>0 }, and then
went to an elementary extension R∗ R satisfying it.
• Consider C as a ring, and study maximal 1-types over Z. There are two
kinds:
– algebraic types, types which contain a formula P (x) = 0 for a poly-
nomial P with coefficients in Z.
– the transcendental type, which contains {P (x) 6= 0 : P ∈ Z[X]}.
• We build on the previous example to study (2-)types of pairs of complex
numbers p(x, y) over Z. There are five kinds:
– both x and y are algebraic over Z: the type boils down to {P1 (x) =
0, P2 (y) = 0} for some polynomials P1 and P2 .
– x is algebraic over Z, but y is transcendental: the type reduces to
{P1 (x) = 0} ∪ {P (y) 6= 0 : P ∈ Z[X]} for a certain polynomial P1 .
– x is transcendental, y is algebraic: p(x, y) = {P (x) 6= 0 : P ∈ Z[X]}∪
{P2 (y) = 0} for some P2 .
– x and y are transcendental but the family (x, y) is algebraically de-
pendent: the type contains {P (x) 6= 0 : P ∈ Z[X]} ∪ {P (y) 6= 0 : P ∈
Z[X]} ∪ {Q1 (x, y) = 0} for some Q1 ∈ Z[X, Y ].
115
– The family (x, y) is algebraically independent: the type is p(x, y) =
{Q(x, y) 6= 0 : Q ∈ Z[X, Y ]}. This is the type of any pair of tran-
scendence degree 2.
One measures richness of a structure by the amount of types it realizes.
Definition 1’.5.12 (ω-saturated). An L-structure M is ω-saturated if any 1-
type over finitely many parameters is realized in M.
Example 1’.5.13.
• A finite structure is always ω-saturated, but this is a trivial example!
• Any DLO is ω-saturated.
• (N, s) is not ω-saturated. Let ϕn (x) be the formula x 6= sn (0). Then the
1-type p(x) = {ϕn (x) : n ∈ N} uses 0 as its only parameter, and is not
realized in N.
• (N, <) is not ω-saturated. Let ψn (x) be the formula ∃x1 . . . ∃xn x1 < · · · <
xn < x. Then the 1-type {ψn (x) : n ∈ N} uses no parameters and is not
realized in N.
Proof . This proof relies on ordinal induction, and may be omitted freely; the
result may not.
We enumerate all finite tuples a of elements of M , and all 1-types over a,
finding an enumeration (by some ordinal λ perhaps bigger than ω) of all 1-
types over all finite tuples of M : {pi (x) : i < λ}. The reader not familiar with
ordinals ought to suppose λ = ω for simplicity; in this case the construction is
very understandable.
Let M0 = M. For each i, pi (x) is a 1-type over Mi , so it is realized in some
elementary extension Mi+1 of Mi (a quick look at Remark 1’.5.10 if necessary).
Take unions at limit ordinals (this is required only if λ > ω). Let M1 = Mλ .
As we have taken the increasing union of an elementary chain, Lemma 1’.1.16
says that M M1 . By construction, all 1-types over finite tuples of M are
realized in M1 .
M1 need not be ω-saturated, as we have added new elements, so we might
have created new types. But we resume the construction, getting M2 . After
countably many steps, the union M∗ is ω-saturated, and an elementary exten-
sion of M by Lemma 1’.1.16 again. e:ultraproductssaturated
116
Remark 1’.5.15. One has no control on the cardinality of an ω-saturated ele-
mentary extension M∗ , even starting with a countable base model M. Indeed,
at each step the number of types might be huge.
For instance, if one considers Z as a ring, then there are continuum many
different 1-types: for each subset I of the primes, consider pI = {x 6= 0} ∪
{∃y x = p.y : p ∈ I} ∪ {∀y x 6= p.y : p 6∈ I} which expresses that the only primes
dividing x are the elements of I.
So there is no countable, ω-saturated elementary extension of the ring Z.
Dealing with 1-types or n-types is the same, as a quick induction reveals.
Lemma 1’.5.16. M is ω-saturated iff any n-type over finitely many parameters
is realized in M.
Proof . Induction. Suppose any n-type with finitely many parameters is real-
ized in M, and let p(x, y) be an n + 1-type with finitely many parameters a (x
still stands for (x1 , . . . , xn ); the n+1th variable is y).
Let q(x) = {∃y ϕ(x, y, a) : ϕ(x, y, a) ∈ p}. This is clearly an n-type over a;
by induction, it is realized by an n-uple b ∈ M .
We now let r(y) = {ϕ(b, y, a) : ϕ(x, y, a) ∈ p}. This is clearly a 1-type over
b, a, which is again a finite tuple. So r is realized in M: there is c ∈ M realizing
it. It is clear that b, c realizes p.
Remark 1’.5.17. Observe that one couldn’t have taken both projections of p,
constructing and realizing first q, then s(y) = {∃x ϕ(x, y, a) : ϕ(x, y, a) ∈ p}.
Indeed, even though c is a realization of s, (a, c) has no reason to realize p!
Again, why are ω-saturated models so important? Because when one tries to
check completeness of a theory, one may focus on them, which are presumably
nicer than the general models. Here is a generalization of Lemma 1’.1.6 to
weaker assumptions.
Lemma 1’.5.18. T is complete iff any two ω-saturated models are elementary
equivalent.
Proof . Let M, N be models of T ; we show that they are elementarily equiva-
lent. By Theorem 1’.5.14, there are M∗ M, N ∗ N which are ω-saturated.
As M ≡ M∗ and N ≡ N ∗ , M∗ and N ∗ are models of T . By assumption, they
are elementary equivalent. It follows M ≡ M∗ ≡ N ∗ ≡ N ; we are done.
Here is our generalization of Corollary 1’.5.7.
Theorem 1’.5.19. Let T be a first-order theory such that for any ω-saturated
models M, N |= T , local isomorphisms between M and N form a Karp family.
Then T eliminates quantifiers.
Proof . We use the elimination theorem, Theorem 1’.5.4. Let M, N |= T ; we
let m ∈ M , n ∈ N satisfy same quantifier-free formulas; we want to show that
they satisfy the same formulas.
117
First, we may assume that M and N are ω-saturated. By Theorem 1’.5.14,
there are ω-saturated elementary extensions M M∗ and N N ∗ . As m
satisfies the same formulas in M and M∗ , it actually suffices to work in M∗
and N ∗ , that is we may assume M and N are ω-saturated.
As m and n satisfy the same quantifier-free formulas, they are locally iso-
morphic, so by assumption they are ∞-isomorphic. They must therefore satisfy
the same formulas. Exercise 1’.10
Lecture 26 (Examples)
1’.5.3 Examples
We eventually shed light on all notions by studying examples.
DLO’s
The theory of dense linear ordering has already been defined and studied in
§1’.4.1. We put the existing pieces together.
Proposition 1’.5.20 (DLO’s). DLO is complete and decidable; it eliminates
quantifiers. The theory is ℵ0 -categorical but not ℵ1 -categorical. Moreover, any
model is ω-saturated and local isomorphisms are ∞-isomorphisms.
Proof . We know everything, but here is how a study could go. Past the two
first, many steps are permutable.
(i). The first step is to realize that any DLO is indeed ω-saturated.
(ii). Then, prove that any local isomorphism is an ∞-isomorphism.
(iii). Moreover, as there are no constants, ∅ is a local isomorphism. It follows
from (ii) that any two models are ∞-isomorphic.
(iv). As a consequence of (iii) and with Proposition 1’.4.14, any two countable
models are isomorphic; the theory is ℵ0 -categorical.
(v). One may uses the Łoś-Vaught criterion (Corollary 1’.3.14 to argue that
the theory is complete.
(vi). It would be as simple to say that any two models are elementary equivalent
by (iii); so the theory is complete, again.
(vii). As the theory is clearly effectively presented (finitely many axioms!), and
complete, it is decidable (Corollary 1.6.3).
118
(viii). Turning back to categoricity questions, one would see that the theory
is not ℵ1 -categorical (Counter-example 1’.4.15). It follows from Morley’s
theorem (Theorem 1’.2.16) that it is not κ-categorical in any κ ≥ ℵ1 .
(ix). It follows from (ii) that DLO eliminates quantifiers (Theorem 1’.5.19).
(i). A model is a set equipped with a function s which is onto except on 0, and
which has no cycles. There are therefore two kinds of orbits: the orbit of
0, which starts and never ends, and other orbits, which have no starting
nor ending point: they are like (Z, s).
``
We see that the general model of T has the form Mi = N I Z, where
I is any set. There is no way to order Z-like orbits.
(ii). Local isomorphisms are easily understood; one would however fail to es-
tablish
` back-and-forth between any two models. It is clear that N and
N Z are not ∞-isomorphic (being countable they would be isomorphic
by Proposition 1’.4.14).
(iii). In particular, the theory is not ℵ0 -categorical. On the other hand since
isomorphism classes are determined by the sole set I, which bears no
structure, T is obviously κ-categorical in any uncountable κ.
(iv). One may apply the Łoś-Vaught criterion to derive completeness of T (and
its decidability, as it is effectively axiomatized).
However we insist that the sooner one understands the ω-saturated models,
the better.
(v). The 1-type p(x) = {sn (0) 6= x : n ∈ N} (no parameters! ` N appears as
indexing set!) is clearly not realized in N, but it is in N Z. On the other
hand, if we fix a ∈ Z, then the 1-type q(x) = {sn (a) 6= x : n ∈ N} over a is
119
` ` ` 0
not realized in N Z. If we want to realize it, we must go to N Z Z.
And so on.
``
We then understand that a model is ω-saturated iff of the form N I Z
for an infinite set I.
(vi). One then sees that local isomorphisms between ω-saturated models do
form a Karp family: suppose that a '0 b are drawn from ω-saturated
models, and let α be on the same side as a. We want to find β such that
a, α '0 b, β.
• If α lies in the orbit of some ai , say α = sn (ai ) for n ∈ N, we let
β = sn (bi ). As a '0 b, this creates no confusion!
• If on the other hand α lies in a new orbit, we pick β in a new orbit
too: this is possible, as I is infinite!
(vii). Any two models are clearly locally isomorphic (the orbit of 0 is similar
in either model). It follows from (vi) that any two ω-saturated models
are ∞-isomorphic. It follows that the theory is complete and eliminates
quantifiers.
(viii). In particular, any inclusion of structures is elementary. As N embeds into
any model, it embeds elementarily into any model.
The Ordering
Proposition 1’.5.22 (N, 0, <). Let T be the theory:
• ∀x ¬(x < x)
120
Proof . Things are a little more subtle here.
(i). A key observation is that this theory interprets (N, 0, s): one can define (by
` ` function s. In particular, any model
a first-order formula) the successor
of T has the form MI = N < (I,<) Z, where I is now an ordered set.
(ii). One must be very careful that local isomorphisms do not form a Karp
family between ω-saturated models. Indeed, pairs (x, x+ ) and (x, x++ )
are locally isomorphic, but back-and-forth between them is not possible!
We already have a feeling that distance comes into play. But we reserve
language-related issues for the end.
(iii). If one has studied ω-isomorphisms (§1’.4.3), one sees that any two models
are ω-isomorphic, so the theory is complete. There is another proof in (v)
but this one is recommended, being simple and elegant.
(iv). For a model to be ω-saturated, I must be infinite. This is however
not enough. Fix a in some copy of Z, and consider the 1-type p(x) =
{∃x1 . . . ∃xn−1 a < x1 · · · < xn−1 < x : n ∈ N} over a. It means that x
is infinitely bigger than a, that x lies in a copy of Z lying above that in
which a lives.
In particular, we understand that for the model to be ω-saturated, I must
have no largest element. And similarly, it must have no least element: it
has no endpoints. This is still not enough.
Fix a < b in I, and consider the 1-type q(x) = {∃x1 . . . ∃xn−1 a < x1 · · · <
xn−1 < x : n ∈ N} ∪ {∃x1 . . . ∃xn−1 x < x1 · · · < xn−1 < b : n ∈ N} over
(a, b): it expresses that the orbit of x contains neither a nor b, that x lives
in a copy of Z which lies between that of a and that of b. This is density
of the ordering I!
And we now see that MI is ω-saturated iff I is a DLO.
(v). One could reprove completeness: the theory is of course not ℵ0 -categorical
(neither is it ℵ1 -categorical, since DLO isn’t), but any two countable, ω-
saturated models are isomorphic.
Indeed, let MI and MJ be such models. I and J are countable DLO’s,
whence isomorphic. We then construct an isomorphism MI ' MJ by
identifying copies of Z.
In particular, any two countable and ω-saturated models are elementary
equivalent. This is however not enough. Let N be any model; by the
Löwenheim-Skolem theorem (Theorem 1’.3.13), it contains a countable
elementary substructure N0 = MI , where I is countable. Now I embeds
into a countable DLO I 0 , so N0 embeds elementarily into N1 = MI 0 . It
follows that N ' N1 , and completeness is proved.
This method is however not recommended, as it relies on the existence of
countable, ω-saturated models; one should prefer the proof of (iii).
121
(vi). We let L = {0, <} and L0 = L ∪ {dn : n ∈ N} be the expanded language.
Notice that WFF(L) = WFF(L0 ); we don’t change the expressive strength
of the language, but merely force certain quantified formulas to now be
quantfier-free.
In particular, a model of T is ω-saturated in L iff ω-saturated in L0 : our
description of ω-saturation does not change.
(vii). One now sees that L0 -local isomorphisms between ω-saturated models are
L0 -∞-isomorphisms.
From there, quantifier elimination in L0 is immediate. As N ⊆ MI is also
an inclusion of L0 -structures, one finds N MI .
122
Corollary 1’.5.26 (Hilbert’s Nullstellensatz). Let K be an algebraically closed
field, X = (X1 , . . . , Xn ) a tuple of indeterminates, and I C K[X] an ideal of
K[X]. Then I has a solution in Kn .
Proof . We may assume that I is a maximal ideal m; notice that by Noetherian-
ity, m is finitely generated. In particular, existence of a solution is a first-order
formula with parameters in K.
We let L be the algebraic closure of the field K[X]/m; clearly K ⊆ L. As
both are models of ACF and the theory eliminates quantifiers, one actually has
K L. Now there is a solution to m in Ln (the class modulo m of the tuple X);
since this is first-order, there is a solution in Kn too.
More interesting results hold, but one unfortunately has to stop some day.
1’.6 Exercises
Exercise 1’.1. Show Lemma 1’.1.16, that is show that given an increasing
elementary chain (Mi )i∈I of L-structures, the union M∗ is an elementary ex-
tension of each Mi .
Exercise 1’.2. Let T be a theory which has models of cardinality n for arbi-
trarily large n ∈ N. Show that T has an infinite model.
Exercise 1’.3. Finish the proof of Lemma 1’.1.11 by proving directly that when-
ever M ⊆ N and ϕ(m) is a universal formula with parameters m in M , one has
N |= ϕ(m) ⇒ M |= ϕ(m).
Exercise 1’.4. Let L = {R}, a binary relation, and T express that R is an
equivalence relation.
1. Let M, N |= T . Find a necessary and sufficient condition for M |= N .
2. Let M ⊆ N |= T . Find a necessary and sufficient condition for M N .
No justification is required.
Exercise 1’.5. Let H, G be groups with H G (the language contains Lgrp =
(e,−1 , ·), but could be larger).
1. Prove that if G is simple, so is H.
2. Prove that the converse may fail.
Exercise 1’.6. A theory T is universal if it the set of consquences of a set of
universal sentences.
Also, if T is a theory, let T∀ be the set of all its consequences that are
universal; that is
123
1. The language is {·}. Is the theory of groups universal? (no justification
required)
2. The language is {e,−1 , ·}. Is the theory of groups universal? (no justifi-
cation required)
3. Show that T is universal iff T∀ |= T .
Exercise 1’.7. Two theories T and T 0 are companions if every model of T is
included (not necessarily elementarily!) in a model of T 0 , and vice-versa.
1. Find an interesting companion of the theory of orderings. Find an inter-
esting companion of the theory of rings.
2. Prove that T and T∀ are companions.
3. Deduce that T and T 0 are companions if and only if T∀ and T∀0 are com-
panions.
4. Prove that T is universal if and only if every substructure of a model of
T is a model of T .
5. Prove that T∀ ⊆ T∀0 if and only if every model of T 0 is included (not
necessarily elementarily, well-understood) in a model of T .
Exercise 1’.8.
1. Suppose M1 ≡ M2 . Show that there is M∗ such that M1 and M2 embed
into M∗ elementarily.
2. Let N be a model which embeds elementarily into Mi for each i ∈ I.
Show that there is a common elementary extension M∗ Mi in which
all copies of N are glued togehter.
Exercise 1’.9. Let T be a theory and ϕ(x) be a formula. Prove that ϕ is
equivalent modulo T to an existential formula if and only if:
for all models M and N of T with M ⊆ N and a ∈ M ,
one has M |= ϕ(a) ⇒ N |= ϕ(a).
[Hint: one implication is trivial. For the other, study the set Ψ := {ψ(c) an
existential formula such that T |= ψ(c) → ϕ(c)}. Prove that T1 := T ∪ {¬ψ(c) :
ψ ∈ Ψ} ∪ {ϕ(c)} is not consistent.]
Exercise 1’.10. Show the converse of Theorem 1’.5.19, that is show: if T elim-
inates quantifiers and M, N are ω-saturated models of T , then local isomor-
phisms form an ∞-isomorphism family.
Exercise 1’.11. The language consists of a single binary relation symbol −.
The theory of graphs is given by the axioms:
• ∀x ¬(x − x)
124
• ∀x∀y (x − y → y − x).
A graph G is connected if for any x, y ∈ G, there is a path from x to y. (This
definition does not sound first-order: we shall prove it.) Show that there is no
theory whose models are exactly all connected graphs.
Exercise 1’.12. Consider the two groups Q and Q ⊕ Q. How isomorphic are
they (as groups)? Same question when they are considered as rings.
Exercise 1’.13. We consider the following <-structures:
a a a
A = ... Z ... and B = A B
< < <
| {z }
Z
`
where X < Y means that the disjoint union is ordered in such a way that every
y ∈ Y lies above every x ∈ X.
How isomorphic are A and B?
125
4. Prove that T is universal if and only if every sub-structure of a model of
T is a model of T .
5. Prove that T∀ ⊆ T∀0 if and only if every model of T 0 is included (not
necessarily elementarily, well-understood) in a model of T .
1. Explain what the theory says (it has a fairly standard name).
2. Describe countable models of T (no justification required).
Exercise 1’.20. Let L = {B, R, <} where B and R (“blue” and “red”) are unary
predicate symbols, and < is a binary relation symbol. Consider the following
conditions:
• every point is either blue or red, but not both
126
• between any two points, there is a point
• between any two blue points, there is a red point
• between any two red points, there is a blue point
1. Write down L-sentences expressing the above axioms. Let T be the re-
sulting L-theory.
2. Show that an order-preserving function need not be a local isomorphism.
6. Is T ℵ0 -categorical? Is it ℵ1 -categorical?
7. What happens if we drop the clause that between any two points there is
a point?
127
Index of Notions
128
in First-Order Logic, 51
Satisfiability
in Propositional Logic, 12
in First-Order Logic, 52
ω-Saturation, 116
Sentence, 49
Signature, 46
Skolem Function, 101
Standard Part*, 74
Structure, 50
Substitutable Term, see Term
Substructure, 93
Elementary Substructure, 95
Term, 47
Substitutable Term, 53
Theory
in Propositional Logic, 9
in First-Order Logic, 49
of a First-Order Structure, 92
Truth Assignment, see Assignment
Type, 115
of a Tuple, 109
Ultrafilter*, 75
Ultraproduct*, 76
Weakening, 16
Well-Formed Formula
in Propositional Logic, 8
in First-Order Logic, 48
Witness, 63
129
Index of Results
Decidability*
Semi-D. of a Theory, 33, 70
of a Complete Theory, 33, 71
Generalisation Theorem, 64
Łoś’ Theorem*, 77
Łoś-Vaught Completeness Test, 103
Löwenheim-Skolem Theorem
Descending Version, 102
Ascending Version, 103
Final Version, 103
QE
Criteria, 113, 117
of DLO, 118
of Th(N, 0, <), 120
130
List of Lectures
131