(Cambridge Studies in Linguistics) Marcus Tomalin - Linguistics and The Formal Sciences - The Origins of Generative Grammar-Cambridge University Press (2006)
(Cambridge Studies in Linguistics) Marcus Tomalin - Linguistics and The Formal Sciences - The Origins of Generative Grammar-Cambridge University Press (2006)
MARCUS TOMALIN
Downing College, Cambridge
Cambridge, New York, Melbourne, Madrid, Cape Town, Singapore, São Paulo
Cambridge University Press has no responsibility for the persistence or accuracy of s
for external or third-party internet websites referred to in this publication, and does not
guarantee that any content on such websites is, or will remain, accurate or appropriate.
since feeling is first
who pays any attention
to the syntax of things
will never wholly kiss you;
e. e. cummings
Contents
Acknowledgments page xi
List of mathematical symbols xii
List of abbreviations xiii
1 Introduction 1
2 The consequences of analysis 21
2.1 Chapter overview 21
2.2 Calculus: doubts and disputes 22
2.3 Rigour, arithmetic, and axioms 26
2.4 Set theory and paradoxes 29
2.5 Logicism 32
2.6 Formalism 38
2.7 Intuitionism 45
2.8 Evangelism and pedagogy 47
3 Mathematical linguistics 54
3.1 Chapter overview 54
3.2 Axiomatics 55
3.3 Recursive definitions 60
3.4 Logical systems 67
3.5 Constructional system theory 73
3.6 Constructive nominalism 84
3.7 Formal linguistic theory 88
3.8 New directions 106
ix
x Contents
6 Conclusion 183
Notes 201
Bibliography 209
Index 220
Acknowledgments
‘∀’ indicates universal quantification: ‘∀x[φ(x)]’ means ‘for all x it is the case
that φ(x) holds’.
‘∃’ indicates existential quantification: ‘∃x[φ(x)]’ means ‘there exists an x for
which φ(x) holds’.
‘¬’ indicates negation: ‘¬φ(x)’ means ‘it is not the case that φ(x) holds’.
‘→’ indicates implication: ‘ p → q’ means ‘if p, then q’.
‘∧’ indicates conjunction: ‘ p ∧ q’ means ‘ p and q’.
‘∨’ indicates disjunction: ‘ p ∨ q’ means ‘ p or q’.
‘N’ indicates the set of natural numbers: N = {1, 2, 3, ...}.
‘∈’ indicates ‘is a member of’: ‘x ∈ A’ means ‘x is a member of set A’.
‘∈’
/ indicates ‘is not a member of’: ‘x ∈
/ A’ means ‘x is not a member of set
A’.
xii
Abbreviations
Books/Book-length Manuscripts/Theses
LCW Carnap, R. (1928), Der logische Aufbau der Welt [The Logical
Construction of the World]
LPV Quine, W. V. O. (1953), From a Logical Point of View
LSL Carnap, R. (1937[1934]), The Logical Syntax of Language
LSLT Chomsky, N. (1975[1955]), The Logical Structure of Linguistic
Theory
MMH Chomsky, N. (1979b[1951]), Morphophonemics of Modern Hebrew
MP Chomsky, N. (1995), The Minimalist Program
MSL Harris, Z. S. (1951), Methods in Structural Linguistics
PM Whitehead, A. N. and Russell, B. A. W. (1925[1910]), Principia
Mathematica [The Principles of Mathematics]
SA Goodman, N. (1951), The Structure of Appearance
SS Chomsky, N. (1957), Syntactic Structures
Papers
xiii
1 Introduction
1
2 Introduction
by experts in those fields. It is not a mere reorganisation of the data into a new
kind of library catalogue, nor another speculative philosophy about the nature
of Man and Language, but rather a rigorous explication of our intuitions about
our language in terms of an overt axiom system, the theorems derivable from
it, explicit results which may be compared with new data and other intuitions,
all based plainly on an overt theory of the internal structure of languages; and
it may well provide an opportunity for the application of explicit measures of
simplicity to decide preference of one form over another form of grammar.
(Lees 1957: 377–378)
Bach gives particular emphasis to the influence of the formal sciences upon the
development of TGG. For instance, he states explicitly that the theory ‘has taken
its inspiration from modern logic’ (Bach 1964: 9), and he goes on to suggest
that mathematics, logic, and linguistics have moved closer together during the
last one hundred years (i.e., 1864–1964):
In the last century a great deal has been learned about the structure of deductive
systems (systems of logic, mathematics, axiom systems for various sciences).
Logicians and mathematicians have been concerned more and more with
studying various ‘language systems’ or ‘calculi’ from an abstract point of
view. At the same time, modern linguistics has tended towards describing lan-
guages as abstract formalised systems. In many ways, the theory of language
presented here may be considered the result of a convergence between these
two currents. The grammars that we shall study are attempts to state the prin-
ciples by which sentences of a language may be constructed, in much the same
way that a formalised mathematical theory may be used to construct theorems.
(Bach 1964: 9–10)
Like Lees before him, therefore, Bach states clearly that TGG makes use of
the same kind of axiomatic-deductive systems utilised by various branches
of mathematics, and, in addition, he suggests that linguistics and mathematics
have been converging for at least a century. Later, he observes more specifically
that the use of techniques derived from ‘modern logic and mathematics’ in
TGG may constitute ‘the most lasting result of the linguistic research of the last
decade’ (Bach 1964: 143). Unfortunately, though, Bach feels that ‘to document
this bit of cultural history in detail would take us well beyond the bounds of
this introduction’ (Bach 1964: 144). Consequently, he leaves this topic largely
undeveloped and does not return to it later in the book.
As the 1960s progressed it became increasingly obvious to the international
linguistics community that TGG was significantly more than an ephemeral fad,
and consequently the issue of the historical roots of the theory began to inspire
more interest. Chomsky himself contributed to this general trend when he pub-
lished Cartesian Linguistics: A Chapter in the History of Rationalist Thought
(1966), a somewhat disingenuous attempt to demonstrate that historical prece-
dents existed for concepts such as deep structure and surface structure – concepts
he was then in the process of elaborating. Specifically, Chomsky declared that
TGG could be viewed as ‘a reawakening of interest in questions that were, in
fact, studied in a serious and fruitful way during the seventeenth, eighteenth and
early nineteenth centuries’ (Chomsky 1966: 1), and he sought to argue his case
by focusing on such works as the Port-Royal Grammar (1660) and by reassess-
ing the work of linguists such as Wilhelm von Humboldt. Although (much to his
6 Introduction
annoyance) Chomsky’s book has never really been taken seriously by linguistic
historiographers, who tend to classify it as a work of ideological propaganda
rather than as an objective historical assessment of the development of syntactic
theory, it certainly seems to have inspired an interest in the task of situating
TGG securely within the history of ideas.2 However, since Chomsky was pri-
marily concerned with linguistic research that had been completed before the
mid-nineteenth century, he did not consider the actual development of TGG
itself, and therefore Cartesian Linguistics offers few insights into the emer-
gence of generative grammar. Curiously, a more revealing contemporaneous
insight into the nature of this development can be gleaned from the writings of
the post-Bloomfieldian Charles Hockett (1916–2000), who, by the late 1960s,
had become a rather lonely figure in the world of international linguistics. Dur-
ing the previous decade Hockett had been generally recognised as one of the
dominant linguists of his generation, but, as the years passed, linguistic the-
ory started to take a distinctly Chomskyan turn, and Hockett found himself
marginalised; a prematurely redundant relic of a bygone era. His response was
to publish The State of the Art (1968), a book in which he expressed his dis-
satisfaction with Chomsky’s general approach to syntactic theory, and it was
while he was preparing the ground for his robust critique of Chomskyan syntax
that Hockett recalled the circumstances surrounding the development of TGG
in the 1950s:
Lacking any explicit guidance as to where to turn for a broadened basis for
linguistic theory, Chomsky was forced on his own resources and tastes, and
turned towards the abstract fields of logic, mathematics, and philosophy, rather
than to science. If Harris’ work suggested either of these directions, it was the
former. Indeed, a number of us at the time, in our search for ‘rigor’, were gazing
longingly towards mathematics . . . the move was reasonable, since linguistics
(or language) surely has interconnections of various sorts with these scholarly
endeavours just as it has with anthropology, psychology and biology. (Hockett
1968: 36)
that particular methodologies employed by the formal sciences (i.e., for Hock-
ett, ‘abstract’ sciences such as logic and mathematics) could enable linguistic
theory to be endowed with greater rigour. As will be shown in section 2.3, the use
of the term ‘rigour’ in this context is replete with significance. Unfortunately,
though, like Bach before him, Hockett does not provide a detailed account of
precisely which branches of mathematics he considered to have been especially
influential, nor does he indicate how the linguists managed to acquire a working
knowledge of contemporaneous developments in mathematics.
During the 1970s, as it became increasingly clear that Chomsky’s place in
the history of linguistics was secure, his early work began to be scrutinised
more closely by linguistic historiographers. For instance, John Lyons discussed
various aspects of TGG in his relatively non-technical book Chomsky (Lyons
1970), and while he certainly argues in this text that Chomsky’s work had
ushered in a new period of linguistic research (describing Syntactic Structures
as a ‘short but epoch-making book’ (Lyons 1970: 36)), Lyons is keen also to
emphasise the similarities between TGG and the type of grammatical research
conducted by the post-Bloomfieldians. For example, at one point he remarks that
This revealing summary suggests that during the early 1950s Chomsky
perceived a close association between the methodology of constructional
system theory and the techniques employed by the post-Bloomfieldians,
especially Harris. Unfortunately, as will be shown below, Chomsky’s reference
to ‘Goodman’s ongoing critique of induction’ seems to have convinced certain
linguistic historiographers that Goodman was a rationalist who rejected
empirical procedures when, actually, the opposite was in fact the case. Indeed,
as is shown in section 5.3 below, Goodman’s apparent critique of induction
was merely the prelude to a robust defence of the same, and Chomsky himself
was well aware of this. Quine, who started out as an adherent of Carnapian
logical empiricism, shifted his position more than Goodman, so that by the
1950s he was certainly disillusioned with Carnap’s approach, and, as indicated
in the above passage, his writing from this period influenced Chomsky
directly. Chomsky’s recollections also indicate the direct association between
Goodman’s ideas concerning the simplicity of constructional systems and
his own early preoccupation with the concept of grammatical simplicity. It
should be noted, though, that, while it is illuminating to read Chomsky’s own
reflections upon these aspects of his early work, his reminiscences do not
really reveal anything that was not already implied by certain comments and
footnotes in his original papers and manuscripts.
Another significant passage in Chomsky’s 1975 recollections occurs when
he considers some of the particular branches of mathematics that influenced his
thinking during the formative years in the early 1950s.
of the papers and essays contained in the volume, this false impression is never
corrected.
Another 1994 publication that contains a detailed consideration of the genesis
of TGG is Stephen Murray’s Theory Groups and the Study of Language in North
America: A Social History. While primarily a work of anthropology rather than
a straightforward historiographical study, Murray’s book extends his research
into the development of syntactic theory in the twentieth century that he had
begun in the early 1980s. With specific reference to the sections of the book
that deal with the period 1951–1957, Murray’s research once again involves
digging up forgotten letters and documents that provide insights into the actual
reception of TGG in the 1950s. However, although he robustly renews his attack
on Newmeyeresque interpretations of linguistic history, he does not provide an
account of the relationship between linguistics and the formal sciences during
the 1950s, though the book does emphasise the closeness of the connections that
existed between the post-Bloomfieldians and the proto-generativists, reinforc-
ing the idea that TGG was largely a continuation of existing research concerns.
Indeed, perhaps Murray’s most stimulating and thought-provoking contribution
in his book stems from his interest in generative grammar as a scientific theory.
Influenced by the well-known work of Thomas Kuhn (1922–1996), particularly
The Structure of Scientific Revolutions (1962), Murray discusses the status of
generative grammar as a scientific theory, and while happy to acknowledge that
TGG does appear to meet some of the criteria for an authentic scientific revolu-
tion, he suggests that later versions of the theory fall alarmingly short. Indeed,
the sketch that Murray provides of MIT-based research into generative grammar
in the post-1980 period is remarkable for its condemnatory tone. Describing
Chomsky as ‘an ageing dictator’, Murray summarises the way in which syco-
phantic ‘cohorts’ compete for favour and seek to emulate ‘Chomsky’s contemp-
tuous rhetoric’, and he condemns the general ‘MIT ethos of imperviousness to
criticism’ (Murray 1994: 445). In Murray’s opinion, the main consequence of
this is that linguistics at MIT is more akin to a dictatorship than a scientific
research centre, and, since the influences of the formal sciences upon syntactic
theory is a central theme of this book, some of these issues are of relevance and
will be discussed again later.
Partly prompted by the new revisionist accounts of linguistic history, such as
those by Matthews and Murray, that had appeared in the early 1990s, Newmeyer
collected together a number of his most recent essays and reviews and published
them in 1996 as Generative Linguistics: A Historical Perspective. Clearly, gen-
eral academic interest in linguistic historiography had increased since 1980, for,
in 1996, Newmeyer felt able to claim (somewhat solecistically) that ‘the origin
Introduction 17
and development of generative grammar are now a hot topic’ (Newmeyer 1996:
1). However, the chapter of Newmeyer’s new book that dealt primarily with the
question of origins largely restated his well-known interpretation of the history
of TGG in terms of a Chomskyan revolution, and his main purpose appears to
have been to refute the recent criticism of his earlier work. Compared to the
studies due to Matthews, Huck, Goldsmith, and Murray, Newmeyer’s arguments
are often disappointingly assertive and superficial. In chapter 2, for instance,
he attempts to provide a thorough account of the development of TGG, but the
discussion contains numerous obscurities and errors. For instance, he claims
that Chomsky’s 1951 Master’s thesis was ‘the first to point out that the proce-
dures of American descriptivist linguists can be likened to the programme of
Carnap’s Der logische Aufbau der Welt’ (Newmeyer 1996: 15), yet Carnap’s
work is not explicitly mentioned in either the submitted or published version
of Chomsky’s thesis. Consequently, Newmeyer’s claims are misleading, and
they sometimes seem designed to support a simplistic, pre-existing interpreta-
tion of linguistic history rather than to constitute an intense investigation of the
bewildering complexity of actual events.
As the 1990s approached their conclusion, a number of significant contri-
butions to TGG historiography appeared. For instance, in 1999 a two-volume
collection of papers by various scholars was collected together and published
in honour of E. F. K. Koerner as The Emergence of the Modern Language
Sciences. This selection contains a number of papers that specifically discuss
the development of TGG. One of the most stimulating is Danny Steinberg’s
elaborately entitled ‘How the Anti-Mentalist Skeletons in Chomsky’s Closet
Make Psychological Fiction of his Grammars’. As the title leads one to expect,
Steinberg’s basic argument is that during the 1950s Chomsky was ‘a fervent
formalist and anti-mentalist’ (Steinberg 1999: 267), who, although he began
to shift towards a rationalist stance after 1959 (the basic shift that Lyons had
observed in 1970), never managed entirely to relinquish the ideological vestiges
of his early empiricism. Astonishingly, although Steinberg is keen to argue that
the pre-1959 Chomsky favoured an anti-mentalist approach to syntactic theory,
he does not even mention the influence of Goodman and Quine’s empiricism
upon Chomsky’s early work, nor does he discuss Chomsky’s use of construc-
tional systems and the empirical assumptions inherent in such an approach to
syntactic analysis.
Since the start of the new millennium a number of studies have been pub-
lished that are concerned either entirely or in part with the origins of TGG.
For instance, Giorgio Graffi’s 2001 publication 200 Years of Syntax: A Critical
Survey contains a whole section (specifically, section 8.4) that considers ‘The
18 Introduction
theory during the first half of the twentieth century. The main purpose is to set
the stage for the detailed consideration of TGG that is contained in chapters 4
and 5. A number of separate but related strands are followed. The main top-
ics considered are the interest of pre-TGG linguists in the axiomatic-deductive
method, the use of recursive definitions in logic and pre-TGG linguistics, the
development of constructional system theory, the advent of constructive nomi-
nalism, the increasing preoccupation of linguists with logical syntax and logical
semantics, and the gradual perception that, during the 1940s and 1950s, linguis-
tics had become more mathematical. By contrast, chapters 4 and 5 constitute a
sustained assessment of the influence of the formal sciences specifically upon
the development of TGG. Initially, the focus is upon Chomsky’s pre-1956 work,
and the main issues considered are his preoccupation with simplicity criteria,
his early belief in the utility of a constructive nominalist approach to syntactic
theory, and his complex attitude towards the use of techniques derived from
logic in linguistic analysis. The discussion then shifts to Chomsky’s work from
the years 1955–1957, and numerous topics are considered, including his nega-
tive appraisal of stochastic techniques, his advocation of evaluation procedures
over discovery procedures, his rejection of the empiricist implications of con-
structional system theory, his use of constructional techniques in his own work,
his early notion of transformation rules, his deployment of recursive rules, and
the axiomatic-deductive character of 1957-style TGG.
Although this introduction has now reached its natural conclusion, it is nec-
essary to make one final point. While such words as ‘historiography’, ‘post-
Bloomfieldians’, and phrases like ‘the 1950s’ have figured prominently in the
preceding paragraphs, just as they will feature conspicuously in the ensuing
chapters, it would be entirely incorrect to assume that this book only deals with
issues that have become quaint and intriguing exhibits in a dusty museum that
seeks to preserve now extinct linguistic theories. On the contrary, the implica-
tions of the issues addressed in this book for contemporary generative grammar
are potentially profound. Ever since the advent of the Minimalist Program (MP)
in the early 1990s, most leading researchers within the field have attempted,
in various ways, to reduce the theory of generative grammar to its essential
elements, rejecting all theoretical constructs that can be reinterpreted in terms
of more fundamental components; and, intriguingly, many of the elements that
have now come to be viewed as essential and irreducible were originally incor-
porated into TGG when they were adapted from the formal sciences in the 1950s.
For instance, such techniques as the axiomatic-deductive method and formal
recursion, as well as general concerns with theory-internal simplicity, were all
associated, in different guises, with the earliest versions of generative grammar,
20 Introduction
and therefore their continued presence within contemporary versions of the the-
ory is of considerable interest. While these issues are discussed at greater length
in the conclusion, it is important here simply to stress the current relevance of
the topics considered in this book. Indeed, perhaps it could be claimed that to
consider the aims and goals of contemporary generative grammar, without first
attempting to comprehend something of the intellectual context out of which
the theory developed, is to labour in a penumbra of ineffectual superficiality.
These words of caution complete this introduction; the requisite apologia for
yet another retelling of an old story has been offered, and a detailed exploration
of the complex relationship between linguistic theory and the formal sciences
can now commence.
2 The consequences of analysis
21
22 The consequences of analysis
not always be immediately obvious how some of the topics assessed in the fol-
lowing sections relate to syntactic theory, the many associations will eventually
emerge.
these ultimate ratios [ultimae rationes] in which quantities disappear are not
actually the ratios of ultimate quantities, but limits [limites] towards which
the ratios of quantities decreasing without limit [sine limite] always converge,
and to which they approach nearer than by any given difference, but actu-
ally never go beyond [nunquam vero transgredi], nor in effect reach until
24 The consequences of analysis
The demonstration of all this will be easy to one who is experienced in these
matters and who considers the fact, until now not sufficiently emphasised,
that d x, dy, dv, dw and dz can be taken as proportional to the momen-
tary differences [differentiis . . . momentaneis proportionales haberi posse],
either increments or decrements, of the corresponding x, y, v, w, z. (Leib-
niz 1863c[1684]: 223)
First, every concept of the subject had to be explicitly defined in terms of con-
cepts whose nature was held to be already known . . . Second, theorems had
Rigour, arithmetic, and axioms 27
to be proved, with every step in the proof justified by a previously proved the-
orem, by a definition, or by an explicitly stated axiom . . . Third, the definitions
chosen and the theorems proved, had to be sufficiently broad to support the
entire structure of valid results belonging to the subject. (Grabiner 1981: 5)
task of exploring and extending work in this area already largely accomplished
by Dedekind and others.12 However, during the 1870s he began to develop new
ideas, inspired by his number-theoretical investigations, and so began the intel-
lectual odyssey that led him to formulate the rudiments of modern set theory.
Since set theory was largely responsible for triggering the crisis of foundations
which shook mathematics to its core in the early twentieth century, it is necessary
briefly to summarise the basics of the conceptual scheme Cantor introduced.
summary is intended to introduce only those aspects of the theory that relate to
topics considered at length in later sections.13
In the Cantorian tradition a set is viewed as a collection of objects; a given
set constitutes a single entity, and the elements it contains are referred to as
the ‘members’ of that set. For instance, using modern notation, the statement
‘A = {α, β}’ indicates that A is a set that contains two members, the elements α
and β. Sometimes sets can be defined by specifying properties of their members.
For example, the statement ‘C = {x | x is an even number}’ indicates that C is
the set of even numbers, with the symbol ‘|’ being understood to mean ‘such
that’. Set membership is indicated using the symbol ‘∈’ (i.e., ‘α ∈ A’ indicates
that α is a member of the set A), while the symbol ‘∈’ / is defined as the inverse
of ‘∈’ (i.e., ‘γ ∈/ A’ indicates that γ is not a member of the set A). Cantor
introduced the basic operations of set theory that are now referred to as union
(‘∪’), intersection (‘∩’) and difference (a.k.a. complement) (‘\’). In addition,
he defined the notion of a subset: if A and C are both sets, then A is a proper
subset of C if every member of A is also a member of C but it is not the case that
every member of C is also a member of A (i.e., in symbols, ‘A ⊂ C’). Cantor
also introduced the notion of a power set: given a set, A, the power set of A
contains all possible subsets of A and is is denoted symbolically as ‘P(A)’.
Although these various definitions and operations seem innocuous in the finite
realm, one of Cantor’s motivations for developing his theory was to devise a
method of dealing with the infinite sets of points on the real number line. Since
he considered a set to be a self-contained whole (i.e., a collection of individuals
that could be treated as an individual entity itself), he was able to explore the
notion of an infinite set simply by extending the concept of a finite set. For
instance, if A = {x | x is an even number}, then clearly the set A contains an
infinite number of elements, since it contains all the even numbers. Eventually,
following this kind of reasoning, Cantor reached the startling conclusion that
infinite sets are not all the same size. In order to illustrate this idea he used the
notion of one-to-one correspondence. For instance, to give an example of the
correspondence principle being used to indicate that two sets are the same size,
since the elements in the set of whole numbers, I = {1, 2, 3, . . .}, can be put
into one-to-one correspondence with the elements in the set of even numbers,
E = {2, 4, 6, . . .}, these sets can be considered to be the same size, which is
clearly counter-intuitive since E ⊂ I . This result suggests that, in set theory at
least, a part can be equal to the whole.
Cantor’s research into infinite sets created the branch of number theory known
as ‘transfinite arithmetic’. As indicated above, his exploration of the basic
set-theoretical formalism caused him to consider the possibility that there were
Set theory and paradoxes 31
infinite sets of different sizes, and, in order to examine this more fully, he intro-
duced transfinite cardinal and ordinal numbers. Given a finite set A, where
A = {α, β}, the cardinality of A is 2 (in modern notation, ‘|A| = 2’), since A
contains two members. The same ideas can be extended to infinite sets and
Cantor introduced the symbol ‘ℵ0 ’ (aleph-null) to represent the cardinal num-
ber associated with the set of whole numbers (i.e., if I = {1, 2, 3, . . .}, then
|I | = ℵ0 ). By the mid-1890s Cantor had explored various properties of the
transfinite cardinals and, in particular, he was able to demonstrate that the power
set of a given set must have a larger cardinal number than the original set itself,
and this seemingly harmless theorem was to have unforeseen consequences
during the early years of the twentieth century. The transfinite ordinal numbers
(which Cantor also introduced) are similar in principle to transfinite cardinals,
the difference being that they can only be obtained for sets the elements of which
have been organised in a pre-defined sequence. Accordingly, Cantor used the
symbol ‘ω’ to represent the transfinite ordinal associated with the set of whole
numbers. Since, in Cantorian arithmetic, numbers are created by a generative
procedure that involves adding ‘1’ to an existing number, Cantor realised that
an infinity of ordinal numbers could be obtained simply by applying the same
procedure to ω. In this fashion, the infinite sequence
1, 1 + 1, . . ., ω, ω + 1, . . ., ω2 , . . ., ωω , . . . (2.3)
can be produced. Since these ordinals denote the size of corresponding sets, it
follows that an infinity of differently sized infinite sets could be constructed.
Cantor’s Mengenlehre was an undeniably provocative theory, and it quickly
inspired controversy. Nevertheless, it appealed to those mathematicians con-
cerned with the task of arithmetising analysis, since it seemed to offer a plausible
framework that could be used to secure the foundations of number theory once
and for all. Consequently, Cantor’s work was elaborated by other researchers,
and it began to be used as a theoretical basis in many different sub-branches of
mathematics. Without doubt, the most important post-Cantorian extension of
the theory was due to Ernst Zermelo (1871–1953) and Adolf Fraenkel (1891–
1965), who managed to provide an axiomatic foundation for set-theoretic con-
cepts. Indeed, the so-called ‘ZF’ axiom set they provided is still standardly used
as the basis for modern classical set theory, and it has enabled the discipline to
blossom into a highly sophisticated branch of twentieth-century mathematics.14
However, despite the prevalent enthusiasm for set theory that characterised the
mid-1890s, Cantor himself soon began to identify apparent weaknesses in his
work, and the problems seemed to cluster around the transfinite aspects of the
theory. For instance, in a series of letters to Dedekind written in 1899, Cantor
32 The consequences of analysis
considered the logical validity of the set of all sets and argued that, since this
set contains all other sets, its transfinite cardinal must be larger than any other
transfinite cardinal. However, since the set of all subsets of a given set must
have a larger cardinal than the set itself (as he had previously demonstrated), it
follows that there is a larger cardinal number than the largest cardinal number
(Cantor 1937[1899]). This paradox caused him to recommend that the set of all
sets and its associated cardinal number should not be included in discussions of
the subject since the concept was not coherent. Accordingly, he began to speak
(somewhat vaguely) of consistent and inconsistent sets, the latter including
those that result in paradoxes.
Alarming as the discovery of this paradox was, Cantor later realised that
the problems were not confined to transfinite cardinals, and he began to dis-
cover unexpected inconsistencies in the theory of transfinite ordinal numbers.
As mentioned above, a basic theorem of transfinite arithmetic states that the
ordinal number of the set of all ordinal numbers up to and including n is larger
than n, and the ordinal number associated with the set of ordinal numbers
{1, 2, 3, . . ., ω} is ω + 1. Therefore the set of all ordinals should be associated
with an ordinal that is larger than the largest in the set. But this is a contradic-
tion, since the set contains all ordinals. Cantor communicated this concern to
Hilbert in c. 1896, and it later became known as the Burali-Forti Paradox after
Cesare Burali-Forti (1861–1931), who discussed the problem in a later paper.
These problems, perhaps ‘paradoxes’, were a cause of real unease when they
became known to the mathematical community: how could set theory possibly
provide a secure foundation for number theory if it clearly contained inherent
illogicalities itself? Indeed, the difficulties resulting from transfinite arithmetic
largely inspired the various proposals for securing the foundations of mathe-
matics that coalesced into the three dominant ideologies that came to be known
(rather too simplistically perhaps) as Logicism, Formalism, and Intuitionism.
Since these philosophical-mathematical movements ultimately influenced the
development of linguistics in the twentieth century, it is necessary to consider
them separately in some detail.
2.5 Logicism
Although logic has its roots in antiquity, the development of modern symbolic
logic was given significant impetus in the seventeenth century when Leibniz
began to consider the possibility of a universal symbolic language that could
be used to clarify all arguments and resolve all disputes, as proposed in his
Dissertation Concerning the Art of Combinations (1666). Leibniz’s ideas were
Logicism 33
extended only spasmodically by his immediate successors, and it was not until
the start of the nineteenth century that more focused and ambitious treatises
concerning logic began to appear. In England, Augustus De Morgan (1806–
1871), George Boole (1815–1865), and Francis (F. H.) Bradley (1846–1924)
made significant contributions to the logic of relations and explored the con-
nections between logic, algebra, and probability theory; in Germany, Hermann
Grassmann (1809–1877) and Ernst Schröder (1841–1902) advanced the theory
of logical operators and clarified the nature of the identity relation; while, in
America, Charles Peirce (1839–1914) fused the work of Boole and De Mor-
gan, thus creating a Boolean logic of relations which required him to adopt the
basic principles of the propositional calculus.15 However, while these various
research trends served to augment the range and power of symbolic logic, it
was Gottlob Frege (1848–1925) who first proposed that logic could provide a
secure foundation for arithmetic and (by extension) mathematics in general. In
his first work devoted entirely to logic, Concept-Script (i.e., Begriffsschrift),
which was published in 1879, Frege presented a sophisticated logical system
that included the truth-functional propositional calculus, the analysis of propo-
sitions as functions and arguments (rather than the traditional subject-predicate
decomposition), the basic theory of existential and universal quantification, the
use of derivations based entirely on the expression-form of statements, and
numerous other procedures and techniques that have since become a standard
part of classical logic. However, it was not until his next major work, The
Foundations of Arithmetic, that Frege explicitly suggested that the type of log-
ical system he had outlined in his Concept-Script could be used as a basis
for arithmetic. This bold claim, which he went on to explore more thoroughly
in The Fundamentals of Arithmetic (1893), effectively initiated the Logicism
movement that was to dominate research into symbolic logic during the early
twentieth century.16
While Frege was developing his logico-mathematical philosophy during the
1880s, Peano was busy in Italy extending the work of Boole, Grassmann, and
Schröder, thereby establishing his own school of symbolic logic. In his ambi-
tious Principles of Arithmetic (mentioned in section 2.3 above) he argued (like
Frege) that arithmetic could be constructed upon the foundation provided by
logic. To this end, he set about devising a consistent notational system that would
enable him to axiomatise mathematics using a logical symbolic language. In
addition, one of Peano’s most significant contributions was to move logic away
from the purely algebraic tradition that had dominated in the mid-nineteenth
century, and he achieved this (in part) by examining the relationship between
logic and Cantorian set theory.17 Gradually Peano’s work became known to
34 The consequences of analysis
that he and Whitehead were pursuing was to extirpate fully these paradoxes
from set theory. The first clear statement of Russell’s intent was contained in
his Principles of Mathematics, which appeared in 1903. In the introduction to
this text, Russell articulated his ‘general doctrine’ (which he associated with
Leibniz) that ‘all mathematics is deduction by logical principles from logical
principles’ (Russell 1938[1903]: 5), and, in the ensuing text, he went on to
articulate his convictions more precisely:
In accordance with these objectives, during the years 1903–1910 Russell and
Whitehead collaborated and published a series of papers in which they consid-
ered various difficulties attendant upon the task of seeking (i) to discover the
most parsimonious axiomatic logical system, and (ii) to derive the whole of
mathematics from this basis in such a way as to avoid all ‘fallacies’. Although,
as mentioned above, much of their work involved synthesising the research of
their predecessors and contemporaries, they also made numerous significant
theoretical contributions themselves. Of these, perhaps the most controversial
was the ‘theory of logical types’. Whitehead and Russell had observed that the
paradoxes of set theory invariably involved self-reference of one kind or another,
hence their tendency to refer to them as ‘vicious-circle fallacies’. Accordingly,
the theory of logical types was designed to delimit the extent of permissible
self-reference in an attempt to avoid the paradoxes. As they later explained,
An analysis of the paradoxes to be avoided shows that they all result from
a certain kind of vicious circle. The vicious circles in question arise from
supposing that a collection of objects may contain members which can only
be defined by means of the collection as a whole . . . The principle which
enables us to avoid illegitimate totalities may be stated as follows: ‘whatever
includes all of a collection must not be one of the collection’. (Whitehead and
Russell 1925[1910]: 37)
36 The consequences of analysis
Although an arbitrary and rather elaborate ‘principle’, the theory at least pro-
vided a practical way of avoiding paradoxes while developing set-theoretical
concepts from the axioms of a given logical system.19
Whitehead and Russell’s decade of research into the feasibility of their Logi-
cist project resulted in the publication of Principia Mathematica (hereafter
PM) during the years 1910–1913. The influence of PM upon the development
of logic during the twentieth century cannot really be overestimated. Although
it contains inevitable inconsistencies and obscurities, this work still constitutes
the most profound attempt to reduce mathematics to logic. Consequently, it is
necessary to consider the form and content of PM in some detail.20
The various chapters of PM were subdivided in sections which were further
subdivided into ‘numbers’, indicated by an asterix and a numeral (i.e., *1),
which served to facilitate cross-references. In numbers *1–*5, the so-called
‘theory of deduction’ was developed, and this effectively constituted the propo-
sitional calculus. Accordingly variables were introduced to denote elementary
propositions (e.g., p, q) and the fundamental logical operators negation (‘¬’),
disjunction (‘∨’), conjunction (‘∧’), and implication (‘→’) were presented and
defined along with primitive propositions. This latter group includes informal
statements of the type ‘Anything implied by a true proposition is true’, and
these were used to enable more complex propositions to be derived. The logi-
cal operators, listed above, were not all assumed to be primitive. Rather, only
disjunction and negation were defined directly, and the definitions of the other
operators were constructed from these. For instance, using the symbol ‘=df ’ to
denote definition, implication can be defined as
p → q =df ¬ p ∨ q (2.4)
indicating that the statement ‘p implies q’ and the statement ‘either not-p or q’
are functionally equivalent. From this minimal basis, essential non-primitive
propositions were then derived. For example, the ‘principle of tautology’ is
asserted (i.e., ‘’) as
( p ∨ p) → p (2.5)
and this proposition is associated with the symbol ‘Taut’ in order to facilitate
future reference.
Having established the basic propositional calculus in this manner, White-
head and Russell then introduced the predicate calculus, and this was accom-
plished in numbers *9–*14, where the propositions established for the proposi-
tional calculus were simply extended so that they could be applied to functions
taking variables as arguments (i.e., φ(x)). In addition, the universal operator,
Logicism 37
‘∀’, was introduced, and statements of the form ‘∀x[φ(x)]’ were taken to mean
‘For all x it is the case that φ(x) is true’. Using the existing formalisms, the
existential operator was then defined in terms of the universal operator and
negation as follows
∃x[φ(x)] =df ¬[∀x [¬φ(x)]] (2.6)
where, as previously, ‘ = df’ indicates that the definiendum is being defined.
With the basics of the predicate calculus established for 1-place propositional
functions, Whitehead and Russell extended the framework so that n-place func-
tions could be handled, and, having introduced the predicate calculus in this
manner, number *20 prepared the ground for many of the later technical devel-
opments by introducing the calculi of classes and relations. ‘Class’ is the word
that is used in PM for Cantorian sets. This term derives from pre-Cantorian
theories of aggregates, and, in the first years of the 1900s, Whitehead and Rus-
sell began to use it as the English equivalent of Cantor’s term ‘Menge’.21 As
defined in PM, classes are understood to be groups of variables that satisfy
propositional functions of the form ‘φ(x)’. Class membership can therefore be
defined as
x ∈ ẑ[φ(z)] ≡ φ(x) (2.7)
where ‘≡’ indicates bidirectional implication. Here ‘ẑ’ constitutes the class
determined by the function φ, so (2.7) indicates that the statement ‘x is a member
of the class determined by the function φ’ implies the statement ‘φ(x) is true’
(and vice versa). The definition of relation follows on from the function-based
definition of classes since, in PM, relations are understood to specify a ‘class
of couples’. In other words, the variable pair (x, y) is a member of the class of
couples associated with the relation R, so long as the statement ‘x R y’ holds
for x and y. Since any 2-place propositional function determines a relation,
there is a clearly defined connection between functions and relations, and this
connection is captured by the equivalence
R = x̂ ŷφ(x, y) ≡ x R y ≡x,y φ(x, y) (2.8)
Upon this theoretical basis, the calculus of relations was developed in numbers
*23–*38, during which time various essential propositions were offered in con-
venient shorthand notations. For instance, if the relation R implies the relation
S, then the notation
R ⊂ S = x R y →x,y x S y (2.9)
was used to prevent laborious use of unwieldy symbolism.
38 The consequences of analysis
2.6 Formalism
The foundational movement that came to be known as Formalism, and which
came to be viewed as an alternative to Logicism, was associated primarily with
David Hilbert, one of the most influential mathematicians of the late nineteenth
and early twentieth centuries.23 Hilbert’s early work was devoted to a range of
topics in number theory, analysis, and algebra, and his interest in foundational
issues was not signalled until the publication of his Foundations of Geometry
in 1899. In this booklet, Hilbert attempted to provide a viable axiomatic foun-
dation for geometry, just as Russell had attempted to do two years previously.
General dissatisfaction with the existing axiomatic-deductive geometrical sys-
tems (especially Euclid’s Elements) had been provoked during the nineteenth
century by the proliferation of non-Euclidean geometries which substantially
undermined the role of spatial intuition as a means of validating geometri-
cal arguments. By contrast with the classical Euclidean methodology, Hilbert
endeavoured to remove all latent remnants of geometric intuition by exploiting
Formalism 39
This quotation is taken from Hilbert’s 1904 paper ‘Concerning the Foundations
of Logic and Arithmetic’, which is often regarded as the earliest statement of his
Formalist manifesto, and there is no doubt that in this paper Hilbert introduced
several of the key ideas that were to dominate his mature foundational work.
During the 1910s Hilbert was enchanted by PM and started to write more
enthusiastically about logic as a result. In particular, he came to admire the
powerful symbolic language that Whitehead and Russell had developed in
40 The consequences of analysis
I should like to eliminate once and for all the questions regarding the foun-
dations of mathematics, in the form in which they are now posed, by turning
every mathematical proposition into a formula that can be concretely exhibited
and strictly derived, thus recasting mathematical definitions and inferences in
such a way that they are unshakable and yet provide an adequate picture of
the whole science. (Hilbert 1967a[1927]: 464)
This passage clearly indicates that Hilbert’s proof theory involved two related
tasks. First, a procedure was required that enabled ‘every mathematical propo-
sition’ to be converted into a ‘formula’, then it must be demonstrated that the
formulae obtained could be ‘strictly derived’. The first task stipulates that math-
ematical statements must be formalised (i.e., converted into strings of precisely
defined symbols) so that mathematics as a whole can be viewed simply as
Formalism 41
‘an inventory of formulae’ (Hilbert 1967a[1927]: 465), and more will be said
about the process of formalisation later. The second task involves the deriva-
tion of the formulae within a given system. The overriding concern here is
with the nature of the proof techniques that are utilised, hence Hilbert’s use
of the compound noun ‘Beweistheorie’. Obviously, since this task involves the
manipulation of strings of symbols that represent mathematical propositions, it
can be said to be characterised by a certain (not necessarily vicious) circularity:
proof-theoretical mathematical techniques are used to determine the validity of
(suitably encoded) mathematical propositions. It is this apparent self-reference
that caused the second of Hilbert’s tasks to be referred to as ‘metamathematics’;
that is, mathematics about mathematics.
Having delineated his basic intentions at the start of the paper, Hilbert imme-
diately proceeds to introduce the fundamental machinery he requires, and the
three main components he presents are a set of logical operators, a general
proof schema, and a set of axioms. The logical operators are unremarkable,
and they include symbols for implication, conjunction, disjunction, and nega-
tion, as well as universal and existential operators. These are all common to
PM-style systems, though (as shown later) Hilbert adopts definitions for the
last two that differ from those offered in PM. The general proof schema Hilbert
presents, however, is of greater interest. In the paper a mathematical proof is
informally defined (rather unsatisfactorily) as ‘an array that must be given as
such to our perceptual intuition’ (Hilbert 1967a[1927]: 465), and (more help-
fully) it is stated that a proof consists of a sequence of inferences. The specific
proof schema Hilbert presents takes the form
G
G→F
(2.10)
F
The axioms in groups I–IV are referred to as ‘the logical axioms’, while those
in groups V–VI are called ‘mathematical axioms’ since they involve number-
theoretic concepts. Once again, this highlights the difference between Formal-
ism and Logicism: Hilbert assumes that certain mathematical objects, such
as the numeral ‘1’, are pre-theoretical, existing in the intuition as a thought-
object (‘Gedankeding’), while Whitehead and Russell seek to derive even such
basic objects from the principles of logic. The axiom group that demands most
attention is group IV, which contains the -axiom. This axiom is required in
order to enable transfinite arithmetic to be incorporated within the basic proof-
theoretical framework. More practically, it also permits the vague notions of
‘all’ and ‘there exists’ to be defined. As shown above, the axiom takes the form
‘A(a) → A((A))’, where (A) denotes an object for which the proposition
A(a) holds, if it holds for any proposition at all. Consequently, the universal
and existential operators can be defined as follows:
These definitions should be compared with those used by Whitehead and Russell
in PM, and included in (2.6) above. As mentioned previously, Hilbert’s use of
the -axiom (which uses the function (x)) enabled him to construct proofs for
transfinite arithmetic in accordance with the strictures of proof theory. However,
while the -function facilitated the construction of such proofs, Hilbert often
Formalism 43
Statements such as this are not atypical. Hilbert repeatedly emphasised the
contentual nature of the metamathematical aspects of proof theory. For instance,
in a 1922 paper, while providing an overview of proof theory, he observes:
least, proof theory was considerably more than a game involving the manipu-
lation of meaningless symbols. Statements such as the above, with their focus
upon the differences between formalisation and metamathematical analysis,
should be recalled when the nature of Hilbertian Formalism is considered.
A common misconception presents Hilbert as wanting to reduce the whole of
mathematics to a contentless exercise in symbol manipulation that is performed
in accordance with clearly defined rules. From this perspective, in the Formal-
ist game it is the relationship between the strings of symbols that is crucial,
and the meaning either of the symbols themselves or of the strings they form is
deemed to be irrelevant. This misconstrual of Hilbert’s programme is partly due
to the practice of extracting certain of his comments from out of their imme-
diate context. For instance, as mentioned above, part of Hilbert’s contribution
in his Foundations of Geometry was to demonstrate that the meaning of the
geometrical objects he considered need not be accommodated in order to anal-
yse them coherently. In other words, statements about lines, points, and planes
could just as readily be interpreted as statements about arithmetic objects, or,
as Hilbert allegedly put it ‘tables, chairs and beer-mugs!’ (quoted in Grattan-
Guinness 2000: 208). However, this conventional misinterpretation of Hilbert’s
programme is also the result of his distinction between the formalisation process
and the metamathematical process being ignored. On numerous occasions, for
instance, Hilbert emphasised that the task of converting mathematical propo-
sitions into a formal symbolic language was a mechanical procedure that did
not rely upon considerations of meaning. In his 1927 paper, for example, he
states that ‘in my theory, contentual inference is replaced by the manipulation
of signs according to rules’ (Hilbert 1967a[1927]: 467). Although this obser-
vation refers only to the pre-metamathematical stage of analysis, comments
such as these (when extracted out of context) appear to suggest that it is the
formal relationships between strings of symbols that matter, not the meaning
of the strings themselves, even during the metamathematical manipulations
of these strings. It was the (mis)perceived extremity of this emphasis on the
formal properties of mathematical statements that caused Brouwer to refer to
Hilbert’s programme as ‘formalisme’. However, as demonstrated above, Hilbert
was never so extreme in his own brand of Formalism, and this observation has
caused some commentators to recommend the avoidance of the term when dis-
cussing Hilbert, or at least to insist upon an accurate definition.26 However, as
discussed in section 2.8, it was a caricatured version of Hilbert’s original theory
that was popularised throughout North America and Europe during the 1930s
and 1940s, and which ultimately influenced the development of syntactic theory
in the twentieth century.
Intuitionism 45
The fate of Formalism, or, more specifically, of proof theory, is well known.
In 1931 the young Kurt Gödel published an incompleteness theorem which
demonstrated that, if a formal system is strong enough to prove theorems from
basic arithmetic, then there will always be theorems that are true but which
cannot be proved within the system. In other words, Gödel demonstrated that
the criterion of completeness was a chimera, and this proof appeared to invali-
date the Formalist approach to the foundations problem. Nevertheless, despite
Gödel’s results, a number of mathematicians have continued to work within
the general framework of proof theory and, as shown in chapter 3, the philos-
ophy behind the theory has exerted a profound influence over many different
disciplines.
2.7 Intuitionism
The third major foundational movement, which became known as Intuition-
ism, was originally associated with the Dutch mathematician Luitzen Brouwer
(1881–1966), and it was intended to provide a valid alternative both to Logi-
cism and (incipient) Formalism.27 Brouwer signalled his preoccupation with
foundational issues as early as 1907, when he submitted a doctoral thesis on the
subject, Concerning the Foundations of Mathematics. Although he later refined
considerably the ideas presented in this early work, the thesis nevertheless out-
lines the basic preoccupations that remained remarkably constant throughout
his long career. For instance, in chapter 2 of his thesis, Brouwer explicitly rejects
the assumption that mathematical objects and the symbols used to express them
are equivalent, arguing instead that thought and language are largely separate.
So crucial was this assumption to the whole Intuitionist enterprise that Brouwer
was later to refer to it as ‘the first act of Intuitionism’ (Mancosu 1998: 8–9). The
inevitable result of this act was that he came to view mathematics primarily as a
process of isolated mental construction, accomplished by the individual math-
ematician, which must then be (imperfectly) communicated to others by means
of language, either a natural language or, most frequently in the mathematical
literature, some kind of formal symbolic language. Consequently, intuition is
understood to provide the foundation for all of mathematics, and the Logicist
and Formalist programmes, with their various ways of prioritising linguistic
forms, are deemed to be utterly misguided since they consistently fail to recog-
nise this crucial fact.
While Brouwer’s characteristic approach to the foundations problem was
outlined in his doctoral thesis, he seems to have felt that he had to establish
himself as a leading mathematician if his nascent Intuitionist philosophy were
46 The consequences of analysis
the continuum was a primitive concept that could not be constructed from more
elementary entities. In particular, it could certainly not be built up from sets
of discrete points, and therefore was not an arithmetic manifold of real num-
bers as Cantor (and others) had supposed. Consequently, in a series of papers
published in the 1920s, he began to devise an intuitionistic version of set the-
ory based upon a dynamic conception of continuity in which points on the
real number line were characterised as convergent sequences of nested inter-
vals that were generated by the free choice of the individual consciousness. In
order to achieve this ambitious project, Brouwer sometimes extended existing
set-theoretical techniques and sometimes devised his own radical new meth-
ods, many of which outraged various sections of the conservative mathematical
establishment.31
While Brouwer’s labours impressed many, the unwieldy nature of the con-
structive procedures demanded by an unremitting adherence to Intuitionistic
principles dissuaded all but a small coterie of devoted followers (the foremost
of whom was Arend Heyting (1898–1980)) from adopting Brouwer’s meth-
ods as practical tools, and there is no doubt that Brouwer was discouraged
by this lack of general acceptance. His disquiet increased in the 1920s when
the Formalism–Intuitionism debate became a personal feud between himself
and Hilbert, resulting in his dismissal by the latter, in 1928, from the edito-
rial board of the journal Mathematische Annalen (Mathematical Annals). After
these final disputes Brouwer retreated into relative obscurity, from which he
did not emerge until the 1950s, when he toured the world as an itinerant lec-
turer, still preaching the Intuitionist gospel. These lectures seem to have been
received by his audiences with considerable warmth, although he was mainly
viewed as a curious relic, a remarkable remnant of an earlier age when anxieties
about the foundations of mathematics were obsessively debated by professional
mathematicians, logicians, and philosophers alike.
that Hilbert had introduced. Quine’s interest in the syntax of formal languages
is apparent throughout his book; he devoted a whole chapter to the topic, and
his presentation is conspicuously influenced by Formalism. For instance, while
discussing the syntax of formal logical systems, he introduces an alphabet of
primitive symbols and adds, with reference to strings formed from this basic
symbol set, ‘all these characterizations are formal systems, in that they speak
only of the typographical constitution of the expressions in question and do
not refer to the meanings of these expressions’ (Quine 1940: 283), and it is
this emphasis on the form, rather than the content, of symbolic expressions that
reveals the influence of Hilbert’s proof theory (as mediated by Carnap). As will
become apparent below, such observations became the norm in introductory
texts of this kind during the 1940s and 1950s.
Another presentation of the rudiments of symbolic logic appeared when
Church published his Introduction to Mathematical Logic: Part 1 in 1944 – a
text that (re)appeared in a revised and expanded form in 1956. Like Quine,
Church presented the same basic topics which, by the mid-1950s, were rapidly
becoming an incantational mantra (i.e., primitive symbols, variables, quanti-
fiers, propositional calculus, first-order predicate calculus, second-order predi-
cate calculus, and so on), and his Formalist agenda is expressed unambiguously
when he observes that ‘traditionally, (formal) logic is concerned with the anal-
ysis of sentences or of propositions and of proofs with attention to the form in
abstraction from the matter’ (Church 1956: 1). It is revealing that the task of
using logic to analyse a proof could already be described as traditional, since
this indicates the speed with which Hilbert’s metamathematical programme
had been accepted. Church goes on to discuss the specific topic of the syntax of
logical systems in a subsection of his introduction, and he draws a distinction
between natural and formal languages. Like Russell, Church felt that natural
languages were far more complicated than formal (i.e., artificial) languages
since they had ‘evolved over a long period of history to serve practical pur-
poses of facility of communication’ (Church 1956: 3). Consequently, when
Church used the term syntax, he was referring specifically to the syntax of for-
mal languages, rather than natural languages, and he made a further distinction
between ‘elementary’ and ‘theoretical’ syntax. Elementary syntax is concerned
with ‘setting up the logistic system and with the verification of particular well-
formed formulas, axioms, immediate inferences, and proofs’, while theoretical
syntax constitutes ‘the general mathematical theory of a logistic system or sys-
tems and is concerned with all the consequences of their formal structure (in
abstraction from the interpretation)’ (Church 1956: 59). Clearly, this approach is
rooted in Hilbert’s proof theory, since the distinction is essentially that between
50 The consequences of analysis
passages indicate that, like Church’s presentation, Kleene’s text can be viewed
as yet another example of the more extreme Formalist position that came to
characterise pedagogic texts during the 1940s and 1950s.
One other book, also published in 1952, that deserves comment in this context
is Raymond Wilder’s Introduction to the Foundations of Mathematics. While
Church and Kleene were concerned with presenting the technical apparatus of
symbolic logic and proof theory respectively, reserving an historical overview
of the development of these theories only for footnotes and asides, Wilder
wanted to provide a detailed historical survey of the whole foundations crisis.
Consequently, in addition to the expected topics, his text includes chapters on
the axiomatic-deductive method, set theory, number theory along with separate
sections on Logicism, Intuitionism, and Formalism. It is of particular interest
that Wilder’s book was directly inspired by Young’s text, discussed above. As
Wilder explains in his introduction,
In a general way, the idea of the book is similar to that which motivated J.
W. Young’s Fundamental Concepts of Algebra and Geometry, first published
in 1911. In 1932 I discussed with Professor Young the desirability of a book
such as this one; he agreed thoroughly that it was desirable to write it, if only
to have available a book on foundational concepts that will take into account
the great strides that have been made in Foundations since the publication of
his book. (Wilder 1952: vi)
Clearly, then, by the early 1950s there was an interest not only in the math-
ematical techniques that emerged from the foundations crisis, but also in the
cultural and intellectual history of the period, and Wilder’s influential text was
consciously designed to cater for this need.
The texts mentioned so far are all either mathematical or logical textbooks
that attempted to provide insights into fundamental results in logic and meta-
mathematics. However, other publications began to appear during the 1940s and
1950s that emphasised the wider utility of some of the techniques developed
in these areas. For instance, a number of texts attempted to argue that logic
could be used to facilitate the analysis of natural language. Two influential texts
of this type are Hans Reichenbach’s Elements of Symbolic Logic (1947) and
Paul Rosenbloom’s Elements of Mathematical Logic (1950). Reichenbach’s
text is a curious mixture of conventional expository sections that introduce
standard topics, and idiosyncratic sections that are mainly concerned with his
own research interests. The most significant chapter from the viewpoint of
linguistic theory is chapter 7, entitled ‘Analysis of Conversational Language’,
in which Reichenbach outlines his plans for ‘a logistic grammar’, which stems
52 The consequences of analysis
from ‘the desire to connect logic with the natural use of language’ (Reichenbach
1947: vi). Although the details of Reichenbach’s system need not be considered
here, it is important to note that his basic intention was to create a grammar
for the English language based upon propositional functions. For instance, he
observed that the declarative sentence ‘x loves y’ could be represented by
the propositional function ‘ f (x, y)’, where ‘ f ’ corresponds to the verb ‘love’,
and his interest in this sort of function-theoretic grammar was motivated by
a dissatisfaction with traditional subject-predicate analysis. Indeed, Reichen-
bach claimed that the standard subject-predicate analysis ‘does violence to the
structure of the sentence’ (Reichenbach 1947: 252). As an example, he con-
siders the sentence ‘Peter is taller than Paul’ and claims that this structure
is inadequately analysed by classifying ‘Peter’ as subject and the rest as the
predicate, since clearly, in this example, there is some kind of functional sim-
ilarity between ‘Peter’ and ‘Paul’, and therefore this similarity (as well as the
obvious functional differences) should be captured by a sufficiently detailed
analysis. He goes on to claim that a grammar based on propositional functions
like those mentioned above could capture these structural similarities more
accurately than subject-predicate analysis. Although Reichenbach’s work was
never developed extensively by professional linguists, his claim that natural
language and logic are more compatible than was commonly supposed at the
time intrigued mathematically minded linguists during the 1950s, as will be
shown in section 3.4.
While not as radical as Reichenbach’s text, Rosenbloom’s small volume
also implied that the syntax of formal logical systems and the syntax of natural
language are essentially identical. His main discussion of this issue is contained
in chapter 4, entitled ‘The General Syntax of Language’, and the definition of
a language that he offers is certainly sufficiently broad to include both formal
and natural languages.
A language consists of certain signs, and certain strings of these signs. Its
syntax consists of rules for classifying and transforming these strings. The
alphabet of a language consists of certain basic signs, usually in finite number
[sic]. By a string we mean a finite sequence of signs. A string is exhibited by
writing its signs in linear order from left to right. (Rosenbloom 1950: 152)
Once again the Formalist emphasis on inscription (i.e., the writing of finite
sequences of signs) is apparent here, and, in this respect, the definition given
above is not particularly unusual. What is of interest, though, is the way in
which Rosenbloom goes on to discuss the differences between formal and
Evangelism and pedagogy 53
natural languages. For instance, he implies that these two types of language
differ in degree rather than in kind:
As in all natural languages, including Esperanto, the rules of word and sen-
tence formation in English are so complicated and full of irregularities and
exceptions that it is almost impossible to get a general view of the structure
of the language, and to make generally valid statements about the language.
It is for this reason that mathematicians and logicians prefer to work with
languages like L3 [a formal language defined earlier in the text] with very
simple and regular structures. (Rosenbloom 1950: 153)
The implication, then, is that, although English and other natural languages are
bafflingly complex, and although mathematicians and philosophers find it easier
to work with artificially constructed languages, there is no reason in principle
why natural languages cannot be subjected to the same kinds of formal analysis
as their artificial cousins. As will be shown in chapter 3, this is exactly the sort
of approach that certain linguists gradually adopted in the the 1940s and 1950s,
and it is no coincidence that Chomsky used Rosenbloom’s text as one of his
main sources of information concerning formal syntax during the early 1950s.
3 Mathematical linguistics
54
Axiomatics 55
3.2 Axiomatics
As indicated in chapter 2, the debates concerning the foundations of mathemat-
ics that raged during the early 1900s generated considerable interest at the time,
and the details of the various foundational strategies proposed by the main par-
ticipants were discussed in introductory texts such as those by Young, Quine,
Church, Kleene, and Wilder (as summarised in section 2.8), all of which were
accessible to mathematically inclined linguists. Without doubt the most signifi-
cant linguist actually to follow the progress of the foundational debates closely
at the time was Leonard Bloomfield (1887–1949), and the first of Bloomfield’s
publications to reveal the extent of his preoccupation with mathematics was
his 1926 paper ‘A Set of Postulates for the Science of Language’. In this short
paper Bloomfield suggested that linguists should start to use the same basic
axiomatic-deductive method which had transformed the study of arithmetic
and geometry in the nineteenth century. Bloomfield uses the term ‘postulates’
instead of axioms, and, at the start of his paper, he explains why a postulational
approach could benefit linguistics:
The method of postulates (that is, assumptions or axioms) and definitions
is fully adequate to mathematics; as for other sciences, the more complex
their subject-matter, the less amenable they are to this method, since, under
it, every description or historical fact becomes the subject of a new postu-
late . . . Nevertheless, the postulational method can further the study of lan-
guage, because it forces us to state explicitly whatever we assume, to define
our terms, and to decide what things may exist independently and what things
are interdependent. (Bloomfield 1926: 153)
defining our (often undefined) terms’ (Bloomfield 1926: 153). In other words,
when compared to more fully developed formal sciences (such as mathemat-
ics), linguistics appeared to be infested with errors that could be avoided if
an axiomatic-deductive approach were adopted, and, in accordance with this
proposal, Bloomfield introduced a set of postulates that could provide a secure
foundation for the whole of linguistics. The particular postulates he introduced
included definitions and assumptions such as1
This passage indicates that Bloch’s own attempt to provide a set of axioms
for phonemic analysis was partly inspired by Bloomfield’s work, although his
observation that due to ‘recent theoretical discussion’ the exact system Bloom-
field proposed would have to be altered, indicates that he was aware of the
limitations of the 1926 postulate set. As a result, Bloch himself made several
‘changes of detail’ to Bloomfield’s system. For instance, he rejected Bloom-
field’s assumption that phonemes are actually present in sound waves, adopting
instead the (then) contemporary view that phonemes are abstract linguistic
units. Nevertheless, despite these alterations, the basic postulational method-
ology Bloch used was identical to that proposed by Bloomfield. In particular,
Bloch’s postulates are expressed as statements in English, and these are intended
unambiguously to define all phenomena associated with phonemic analysis. For
instance, Bloch’s first postulate and first definition take the form4
1.1 Postulate 1: There are communities of human beings who interact partly
by the use of conventional auditory signs.
1.2 Definition: Such a community is a speech-community.
58 Mathematical linguistics
φ(0) = q (3.1)
φ(y ) = χ (y, φ(y)) (3.2)
1. add(4, 1) = add(3, 1) + 1
2. add(3, 1) = add(2, 1) + 1
3. add(2, 1) = add(1, 1) + 1
4. add(1, 1) = add(0, 1) + 1
5. add(0, 1) = 1
6. add(1, 1) = 1 + 1 = 2
7. add(2, 1) = (1 + 1) + 1 = 3
8. add(3, 1) = ((1 + 1) + 1) + 1 = 4
9. add(4, 1) = (((1 + 1) + 1) + 1) + 1 = 5
In this example, the initial function call, add(4, 1) triggers a sequence of four
recursive function calls which only terminates in line 5 when add(0, 1) is
reached. Once the recursion has been halted in this manner, the values that
are returned by each recursive call are calculated. Equation pairs of the kind
defined in (3.1) and (3.2) were used by Peano in order to define mathematical
induction, and they have always been at the core of recursive function theory.
As indicated in section 2.6, Hilbert’s interest in primitive recursive functions
was primarily due to the fact that they could replace -functions in mathematical
proofs, and in order to appreciate why this might be desirable, it is worth empha-
sising the relationship between recursive functions and proof by induction.
In mathematics an inductive proof attempts to prove a given theorem by
Recursive definitions 63
establishing that the theorem holds for the first case and for the n + 1th case.
Since the proof is obtained for both an initial instance and a general succes-
sor, then the proof is understood to hold for all cases, since any case can be
derived from the initial instance by repeated applications of the inductive step.
Consequently, although an inductive proof is finite, it covers an infinite number
of cases. Primitive recursive functions permit the same type of iterated infer-
ence and, consequently, Hilbert believed that such functions could guarantee
the validity of proofs which previously had to utilise -functions. Accordingly,
the notion of recursion was established at the core of Formalism and, as a result,
it received considerable attention from other mathematicians. Indeed, the rela-
tionship between recursive functions and induction is extremely close, to the
extent that recursion can be viewed as definition by induction (e.g., Kleene
1952: 217). Further, the fact that recursive functions enable finite proof tech-
niques to be employed was noted by many researchers, and this was often cited
as a desirable and characteristic property. For example, Gödel stated explicitly
that ‘[r]ecursive functions have the important property that, for each given set
of values of the arguments, the value of the function can be computed by a
finite procedure’ (Gödel 1986b[1934]: 348). One obvious question that arises
from considerations of this kind can be expressed as follows: which classes of
number-theoretic functions can be defined recursively? The answer to this ques-
tion was developed during the 1930s, primarily by Gödel, and later elaborated
by his successors, especially Kleene. One crucial notion that was advanced as
a result of this research was that of primitive recursion. Formally, a function is
classified as being a primitive recursive function if it is definable by a series of
applications of the following equations:8
φ(x) = x (3.3)
φ(x1 , x2 , . . ., xn ) = q (3.4)
φ(x1 , x2 , . . ., xn ) = xi (3.5)
φ(x1 , x2 , . . ., xn ) = ψ(χ1 (x1 , x2 , . . ., xn ), . . ., χm (x1 , x2 , . . ., xn ))
(3.6)
φ(0) = q (3.7)
φ(y ) = χ(y, φ(y)) (3.8)
φ(0, x2 , . . ., xn ) = ψ(x2 , . . ., xn ) (3.9)
φ(y , x2 , . . ., xn ) = χ(y, φ(y, x2 , . . ., xn ), x2 , . . ., xn ) (3.10)
The equations above define number-theoretic functions when n and m are
positive integers, i is an integer such that 1 ≤ i ≤ n, q is a natural number,
and φ, ψ, χ , χ1 . . . χm are number-theoretic functions that take the indicated
64 Mathematical linguistics
Definition 3.1 can be directly compared with the equation pair (3.1)–(3.2) above,
since, in both cases, the recursion is induced by the second step. Having con-
sidered the basic implications of adopting such definitions in a mathematical
context, Bar-Hillel goes on to suggest that, in the empirical sciences, defini-
tions have often included statements that are actually recursive, even though
this property has not been explicitly acknowledged. In order to illustrate his
point he states that, ‘for a change’ (Bar-Hillel 1953a: 163), he will consider
several examples from linguistics. As implied, this manoeuvre is rather unex-
pected, since recursive definitions had usually been considered in relation to the
formal sciences (especially proof theory), and had not been explicitly utilised
previously in the context of linguistic theory. In his discussion of this subject,
though, Bar-Hillel adopts a purely language-based approach by using English
as a metalanguage and by employing it to analyse his chosen object language,
French. At the beginning of his analysis, he introduces the following definition
(see Bar-Hillel 1953a: 163):12
In this definition, the terms ‘nominal’ and ‘verbal’ can be understood to mean
‘noun phrase’ and ‘verb’ respectively, and it is important to note that this def-
inition seeks to define sentences which can be infinitely long, in terms of sub-
components combined by means of the conjunction ‘et’. Having introduced
66 Mathematical linguistics
Expressed in this fashion, the definition of a sentence now takes the familiar
form of ‘a pair of simultaneous recursive definitions’, and, Bar-Hillel goes
on to claim that it is ‘simple’ to check that a given compound sentence is ‘a
proper French sentence’, since the structure can be iteratively broken down
into smaller units until the basic constituent units (i.e., clauses) are obtained
(Bar-Hillel 1953a: 164). If this can be achieved, then the sentence is ‘proper’,
if not then (presumably) it is improper (i.e., ungrammatical). Once again it is
clear that one of the consequences of a mathematical (in this case, recursive)
approach to syntax is that grammaticality is most naturally expressed in binary
terms: if a sentence can be analysed in accordance with the posited recursive
definition, then it is grammatical, if not, then it is ungrammatical. Bar-Hillel
ends his short paper with a rallying cry which, though expressed in very general
terms, was bound to intrigue any linguists who were even remotely interested
in the relationship between the formal sciences and linguistic theory:
In conclusion, let me say that, in view of the role played by recursive definitions
in concept formation in empirical science, it is the task of the methodologists
to dedicate time and effort to the evaluation of their precise import in different
fields of inquiry and the task of the scientists to become acquainted with the
recent investigations on recursive definitions to a degree, at least, that would
free them from the misconceptions that have so frequently been connected
with their occurrence in disguise. (Bar-Hillel 1953a: 165)
The basic thrust of the above passage is that certain fields of empirical enquiry
were infested with ‘misconceptions’ primarily because they failed to realise
that they were concerned with phenomena that could most easily be defined
recursively. Given the topic of Bar-Hillel’s own discussion – namely, natural
Logical systems 67
become apparent later, only the work of Leśniewski and Ajdukiewicz will be
considered here.13
Leśniewski studied under Tawardowski, and was appointed to the position of
Chair of Philosophy at Warsaw in 1915. His most significant contribution to the
development of logic was the tripartite logical system he devised, and which
was intended to provide a more secure basis for mathematics than the system
contained in PM. The ambitious nature of this work alone is revealing, since
the very fact that Leśniewski attempted such a project indicates that, during the
1920s, the Logicist movement was still very much alive. His main doubts about
the PM system concerned the imperfections inherent in the kind of formal sym-
bolic language it employed, and he was particularly concerned by the dubious
validity of Whitehead and Russell’s assertion operator. For instance, Leśniewski
juxtaposed the proposition ‘ p’ and the assertion ‘ p’ and asked whether the
two propositions have the same meaning or not. Indeed, he even asked whether
the latter was actually a proposition at all (Leśniewski 1992c[1927]: 181–196).
As for the controversial theory of logical types, Leśniewski argued that it was
unclear whether the theory defined an ontological or a semantic hierarchy, and,
besides, he was convinced that the theory was too intricate to serve as a basic,
intuitive part of a logical system (Leśniewski 1992b[1914]: 115–128). Accord-
ingly, motivated by these concerns, Leśniewski gradually developed his own
alternative to PM during the years 1914–1939, and his system was charac-
terised by a tripartite structure, the main components of which he referred to as
‘protothesis’, ‘ontology’, and ‘mereology’.14 Protothesis and ontology, taken
together, provided the logical foundation of the whole system, the former being
a generalised sentential calculus based upon the notion of equivalence, the latter
being a nominal calculus with the function ‘∈’ (i.e., a stylised form of the
that appears at the start of the word ‘` σ τ ί’ [Greek: ‘it is’]) as its primitive term.
The logical core provided by these components was equal in power to that con-
tained in PM. The third component of Leśniewski’s system, mereology, was an
extralogical theory based upon the single primitive non-logical constant ‘Part’.
In essence, mereology was a reformulation of Cantorian set theory in which
set membership was transitive rather than intransitive. For instance, in classical
set theory, given the fact that (i) x ∈ A and (ii) A ∈ B, it does not necessarily
follow that x ∈ B, while in mereology, if (i) and (ii) hold, then x ∈ B follows.
One of the reasons for preferring this kind of set theory was that it
enabled Russell’s paradox to be avoided when deriving set theory from log-
ical axioms without having to resort to elaborate and artificial methods. How-
ever, Leśniewski was also motivated to construct his transitive alternative to
classical set theory by his nominalistic beliefs. These beliefs were manifest
Logical systems 69
in his assertion that only individuals exist and that, therefore, aggregates or
collections of any kind should not be manipulated as independent entities, but
should rather be defined in terms of their individual elements (see Leśniewski
1992c[1927]: 229–230, especially Axiom II). Consequently, classical sets must
be consistently rejected. This extreme stance can be seen as a direct reaction
to the paradoxes of set theory: if the theory as standardly formulated permitted
contradictions, then it must be reformulated, and nominalism, with its focus on
individual entities and its refusal to accept the abstract notion of a set, consti-
tutes an intellectually plausible alternative framework. With specific reference
to formal symbolic languages, Leśniewski’s nominalism caused him to view
them as finite sequences of inscriptions that could be combined in various ways
to produce longer sequences, and these methods largely anticipated the kind of
approach espoused by the architects of constructive nominalism in the 1940s,
as will be shown in sections 3.5 and 3.6.
As indicated above, perceived flaws in PM inspired Leśniewski to devise
a viable alternative, and Whitehead and Russell’s work also provoked
Ajdukiewicz, who was one of Leśniewski’s young contemporaries at War-
saw. Rather than focusing on the Logicist foundational programme, though,
Ajdukiewicz was more concerned with the task of exploring the syntax and
semantics of formal logical languages of the type synthesised by Whitehead
and Russell. In particular, he was interested in the differences that existed
between such languages and natural languages. Therefore, in a series of
papers, starting with ‘On the Meaning of Expressions’ in 1931, he consid-
ered various technical aspects of semantics and syntax. Often in his published
work he chose to focus on well-defined formal languages simply because,
by contrast, natural languages suffered from ‘vagueness and lack of clarity’
(Ajdukiewicz 1978b[1931]: 26). Nevertheless, he was always keen to consider
his conclusions in relation to natural languages. In particular, in much of this
work, Ajdukiewicz repeatedly emphasised the close connection between syn-
tax and semantics, and he sought to specify various sets of ‘meaning-rules’
(Ajdukiewicz 1978a[1934]: 57) for natural language. Since he often advocated
a relativistic view of language, with a pronounced concentration on situational
interpretation, his work in this area can be viewed (to some extent) as a kind of
proto-pragmatics.15
Ajdukiewicz’s main contribution to syntactic theory, though, was ‘Syntactic
Connexion’ (1936), and, at the very start of this paper, he indicated that his inter-
est in the syntax of formal languages was inspired by the paradoxes of set theory
and the Logicist movement that sought to eliminate them: ‘the discovery of the
antinomies, and the method of their resolution, have made problems of linguistic
70 Mathematical linguistics
s s s s n n n n
s, n, , , . . ., , , . . . , , . . ., , , . . . (3.11)
n nn s ss n nn s ss
When this analytic strategy was applied specifically to natural language, each
basic part of speech was associated with one of these indices. For instance, the
definite article ‘the’ was associated with the index ‘n/n’, while a noun such as
‘lilac’ was associated with the index ‘n’. Consequently, in this manner, syntactic
analysis can be accomplished by means of manipulating indices. It is crucial to
note that, although Ajdukiewicz developed his system of syntactic analysis for
formal languages, he clearly indicated that, as far as he was concerned, such
techniques could usefully be applied to natural languages. This is yet another
indication of the fact that, during the 1930s and 1940s, the supposed distance
that was perceived to separate formal and natural languages appeared gradually
to be diminishing.
Logical systems 71
The particular ‘structural linguists’ that Bar-Hillel cites in a footnote are Charles
Fries (1887–1967) and Zellig Harris (1909–1992), both of whom can be broadly
classified as post-Bloomfieldians. It is perhaps unnecessary to stress that Harris’
work in particular was extremely influential during the 1940s and early 1950s.
Indeed, his Methods in Structural Linguistics (hereafter MSL) (1951), with its
emphasis on distributional discovery procedures, had helped to define the basic
methodology of structural linguistics.16 As will be shown in later chapters,
Bar-Hillel’s conviction that techniques derived from the formal sciences could
be combined with methods developed by structural linguists is one that was
voiced frequently, by various researchers, during the 1950s, and some of the
reasons for the perceived association, or compatibility, between mathematical
techniques and the methodology of structural linguists are considered in section
3.7. Suffice it to say here that, for Bar-Hillel, the connection was due to the fact
that certain techniques derived from logic met the structuralist demand for
analytic procedures that functioned in a mechanical fashion. However, in order
fully to appreciate Bar-Hillel’s work, it is crucial also to note that throughout
his paper the emphasis is upon the description of sentential structure and not
primarily upon the type of mechanical discovery procedures sought by Harris
(and others). As Bar-Hillel puts it himself,
We are not interested here in developing a method which a linguist might use to
ARRIVE at an analysis of a linguistic corpus, but only in a new way in which
he could PRESENT the results of his investigations. (Bar-Hillel 1953b: 47)
Whatever his underlying motivations may have been, though, once Bar-
Hillel had indicated how Ajdukiewicz’s approach to syntactic analysis could
be combined with the methodology of structural linguistics, he started to out-
line the basic theory which (in linguistic circles) would later be referred to as
Categorial Grammar. The basic approach involves the definition and manipula-
tion of fundamental syntactic categories. For instance, given the sentence ‘Poor
John sleeps’, following Ajdukiewicz, Bar-Hillel claims that this structure can
be exhaustively analysed using the two categories ‘n’ and ‘s’, which he defines
as
r n: the category of ‘name-like strings’.
r s: the category of sentences.
With these definitions in place, he then introduces the following derived cate-
gories:
r n/[n]: the category of those strings with an n to their right form a
string that belongs to category n.
r s/n: the category of the string that with an n to the left forms a string
belonging to the category s.
Given the above definitions, a categorial analysis of the sentence ‘Poor John
sleeps’ simply requires the sentence to be associated with the index sequence
n s
n (3.12)
[n] n
In which ‘n/[n]’ denotes the adjective ‘poor’, ‘n’ denotes the noun ‘John’, and
‘s/n’ denotes the verb ‘sleeps’. Consequently, the subsequence ‘n/[n]n’ can
be reduced to n, giving the ‘first derivative’ of (3.12)
s
n (3.13)
n
which is turn can be reduced to ‘s’, forming the ‘second and last derivative’ of
the initial index sequence in (3.12). Since the last derivative of the string belongs
to the category s, the sentence can be considered grammatical since it consti-
tutes a legitimate sentence. As Bar-Hillel observes, and as the title of his paper
indicates, the process sketched above is ‘something like ordinary mathematical
multiplication of fractions’ (Bar-Hillel 1953b: 48), and, as suggested previously,
it is the mechanical nature of this process that allows it to be fused with struc-
turalist linguistic methods: using Bar-Hillel’s analytical procedure, sentences
of natural language can be converted into strings of well-defined symbols in
Constructional system theory 73
Phillip Frank (1884–1966) started discussing various ideas concerning the rela-
tionship between logic, mathematics, philosophy, and science. In 1921 Hahn
returned to Vienna as a lecturer and managed to arrange for Moritz Schlick
(1882–1936) to be appointed to the chair of inductive philosophy. Schlick was
a physicist by training, but was preoccupied with various topics in epistemology
and ethics. He began to lead coffee-house discussion groups in the early 1920s,
and the Vienna Circle emerged out of these gatherings. Other key members of
the group were Karl Menger (1902–1985) and Gödel, and (to put it simply)
this disparate group was unified by a common desire to rescue philosophy from
the clutches of the metaphysicians by making it as precise and as exacting as
mathematics or physics. In particular, the members of the Circle sought to pro-
vide a secure foundation for the sciences, to reject metaphysics, and to utilise
logical analysis for the purposes of philosophical enquiry. Consequently, given
Carnap’s interests, it is no surprise that, when he joined the Circle in 1925, he
quickly established himself as one of its defining figures. Indeed, although LCW
was completed before he travelled to Vienna, in later years the book came to be
viewed as one of the seminal texts of the so-called logical positivist movement
that was associated with the Vienna Circle, and, since the text exerted a pro-
found influence over the work of Goodman and Quine (both of whom in turn
influenced Chomsky), it is necessary briefly to summarise the philosophical
project that the book describes.18
Writing in his ‘Intellectual Autobiography’ of 1963, Carnap recalled that
during the 1920s he had made ‘numerous attempts at analysing concepts of
ordinary language relating to things in our environment and their observable
properties and relations, and at constructing definitions of these concepts with
the help of symbolic logic’ (Carnap 1963: 16). It is intriguing that, from the out-
set, Carnap was interested in ‘concepts of ordinary language’, since this research
was the start of his exploration of logical epistemology, which was ultimately
to influence the development of syntactic theory. His general approach was to
employ the logical system developed in PM as a tool for considering ques-
tions of knowledge acquisition. This basic project was certainly influenced by
Russell’s provocative 1914 publication Our Knowledge of the External World,
which had considered the possibility of such an approach to epistemological
questions.19 Specifically, in LCW, Carnap was explicitly concerned with the
task of creating a Konstitutionssystem, or ‘constructional system’, which he
described as follows:
To begin with, a basis must be chosen, a lowest level upon which all others
are founded. Secondly, we must determine the recurrent forms through which
we ascend from one level to the next. Thirdly, we must investigate how the
objects of various types can be constructed through repeated applications
of the ascension forms. The fourth question concerns the over-all form of the
system as it results from the stratified arrangement of the object types. (Carnap
1967[1928]: 47)
Carnap acknowledges that problems (ii), (iii), and (iv) are closely related to
problem (i), since the choice of basis largely determines the form of the objects
created and the manner of their creation, and these two characteristics largely
determine the nature of the system as a whole. Since the technical issues asso-
ciated with this cluster of interconnected problems are of such fundamental
importance to Carnap’s logical-epistemic approach, and since they were also
explored by his most influential successors, it is necessary to discuss some of
them briefly here.
76 Mathematical linguistics
Back at Harvard, Quine was keen to communicate some of the ideas he had
encountered in Europe to his colleagues. Consequently, he lectured on Carnap’s
latest work in November 1934, and in 1935 he led a series of ‘informal seminars’
concerning the LSL of which ‘the participants were a professor, [David] Prall,
an instructor, Henry Leonard, and some graduate students, including Nelson
Goodman, Charles Stevenson and John Cooley’ (Quine 1985: 122).22
Carnap himself was able to join the group in December 1935, since he was
in America to spend the winter in Chicago. During his stay at Harvard he lec-
tured on the LSL, with Goodman and Quine (along with Prall and Leonard)
functioning as his intellectual bodyguards: ‘we moved with Carnap as hench-
men through the metaphysicians’ camp’ (Quine 1985: 122). These interactions
appear to have been mutually rewarding, and, to some extent, they determined
the course of the research undertaken by all the main participants for the next
few years. In particular, Quine became increasingly concerned with technical
difficulties associated with the syntax of formal languages. His syntactic inter-
ests are apparent both in his pedagogic publications from this time (especially
Mathematical Logic (Quine 1940), which was discussed in 2.8) and certain
journal papers. For instance, inspired by the work of Tarski and Hans Her-
mes (b. 1912), Quine devoted a whole paper to an exploration of the role of
concatenation in formal language theory, suggesting that the concatenation
relation could provide a viable foundation for the whole of arithmetic (Quine
1946). As shown in section 5.4, some of these ideas would later resurface
in TGG.
It was during these seminars devoted to the LSL, therefore, that Goodman and
Quine first came into regular contact, and, given their respective philosophical
predilections, it is no surprise that they quickly became friends and collab-
orators. Although two years older than Quine, Goodman’s academic career
advanced more slowly, but, by the time the two men met, his intellectual inter-
ests had already begun to coalesce. Indeed, he was already actively working
with Henry Leonard on a research project that explored aspects of Carnapian
logical epistemology and which would ultimately provide material for his 1941
Ph.D. thesis. Appropriately, Goodman and Leonard first confessed their secret
project to Quine while the three men were travelling back from Cambridge
(where they had just left Carnap) to Baltimore, and Quine’s recollection of the
conversation indicates that the context of the research was clear to him even at
the time:
We talked in our hotel room until four in the morning. They were concerned
with constructing a systematic theory of sense qualities, and their effort had
Constructional system theory 79
Multigrade relations are predicates that take classes of various magnitudes for
their arguments and which cannot therefore be analysed in terms of, say, dyadic
relations which ultimately define the association between primitive elements in
the system. For instance, as Goodman and Leonard demonstrate, the predicate
‘met with’ is a multigrade relation since any number of people could meet with
any number of other people, and the event of meeting cannot accurately be
analysed as the sum of individual meetings which involve only two people. For
these reasons, and partly motivated by what Goodman and Leonard refer to as
‘considerations of economy’ (Goodman and Leonard 1940: 51), philosophers
working within a broadly logical-epistemic framework had been forced to reject
multigrade relations as primitives in their constructional systems. Accordingly,
Goodman and Leonard were keen to encourage the use of such relations by
introducing the ‘part–whole’ relation, ‘x | y’, indicating ‘x and y have no part
in common’, and by demonstrating that this relation can be used to analyse these
relations more intuitively. As indicated previously, this is the aspect of Goodman
and Leonard’s work that Quine recognised as an independent reformulation of
Leśniewski’s mereology. It is of particular interest that Goodman and Leonard
cite ‘consideration of economy’ as one of the reasons why multigrade rela-
tions had been rejected as appropriate predicates for extralogical bases of con-
structional systems, since such considerations of economy were increasingly to
dominate Goodman’s academic research for the next two decades. More specif-
ically, partly motivated by his exploration of the part–whole relation defined in
his first paper, Goodman came to believe that ‘simplicity’ and ‘economy’ were
crucial aspects of system construction. However, he also recognised that these
related notions were currently being used inconsistently, and that their precise
meaning needed to be clarified. Consequently, during the 1940s and 1950s he
endeavoured to provide this clarification.
Although the term ‘simplicity’ has intuitive appeal, Goodman realised that a
formal definition was required if the centrality of basal simplicity in construc-
tional system theory was to be fully acknowledged. He also perceived that the
existing definitions were conspicuously inadequate. The intuitive assumption
underlying his own approach was that a constructional system possessing a sim-
pler, more economical, basis was preferable to an alternative system that utilised
a more complex basis. Initially, the easiest way of determining the relative sim-
plicity of two given bases appeared to involve merely counting the number
of extralogical primitives, and, certainly, with this end in view, Goodman was
keen to develop ways of reducing the number of such primitives. For instance,
in his 1940 paper ‘Elimination of Extra-Logical Postulates’, which was writ-
ten with Quine, Goodman devised a definitional methodology that effected the
Constructional system theory 81
elimination of basal relations. For example, if a given basis contains the primi-
tive transitive relation ‘Pt’, which indicates ‘is a spatial part of’, then ‘x Pt y’
states that ‘x is a spatial part of y’, and this relation can be determined by the
extralogical postulate
∀x∀y∀z[x Pt y ∧ y Pt z → x Pt z] (3.14)
In this case, rather than taking ‘Pt’ as a primitive element in the basis, the more
general relation ‘O’, which indicates spatial overlapping, could be introduced
instead with the result that the transitivity of the relation ‘Pt’ becomes a theorem
rather than a primitive notion:
A number of the issues raised here are worth highlighting. For instance, it is
important to recognise that, for Goodman, simplicity considerations are only
valid if they involve the replacement of one idea by another, and if the replace-
ment is motivated by ‘some special knowledge concerning the ideas involved’.
Consequently, some kind of simplicity criterion is required that would be able
to quantify the degrees of genuine economy associated with the extralogical
primitives in the respective bases of two given constructional systems and,
in his 1949 paper ‘The Logical Simplicity of Predicates’ Goodman proposed
one possible criterion for nominalistic constructional systems. He begins by
introducing a complexity measure that assigns a value to the predicates in the
basis of a given system. Initially, the measure merely computes the complexity
value for an n-place predicate using the formula 2n − 1. For instance, a 2-place
predicate, a 3-place predicate, and a 5-place predicate are associated with the
values 3, 5, and 9 respectively. This complexity measure reveals that a basis
containing a single 5-place predicate is more complex than a basis that contains
a 2-place predicate and a 3-place predicate, since, in this case, the bases have the
complexity values 8 and 9 respectively. Goodman proceeds to consider other
aspects of predicate structure, including the number of joints and segments,
and, ultimately, the complexity value for a given predicate is determined by
computing the value of the initial complexity measure and adding the number
of joints and segments.24 He also emphasises that these numbers are dependent
upon the available information: if it is not known how many joints and segments
are associated with a given predicate, then (obviously) these values cannot be
incorporated into a measure of complexity.
Constructional system theory 83
The motives for seeking economy in the basis of a system are much the same
as the motives for constructing the system itself. A given idea A need be left
as primitive in a system only so long as we have discovered between A and the
other primitives no relationship intimate enough to permit defining A in terms
of them; hence the more the set of primitives can be reduced without becoming
inadequate, the more comprehensively will the system exhibit the network of
interrelationships that comprise its subject-matter. Of course we are often
concerned less with an explicit effort to reduce our basis than with particular
problems as to how to define certain ideas from others. But such special
problems of derivation, such problems of rendering certain ideas eliminable
84 Mathematical linguistics
The ideas promulgated and debated in these meetings clearly influenced the
thinking of all the participants, and Goodman and Quine eventually articulated
their response in their 1947 paper ‘Steps Towards a Constructive Nominalism’
(from henceforth ‘SCN’). The basic purpose of this work was to implement pre-
cisely the type of nominalistic formal language that had been proposed during
the Harvard discussions, and this project was pursued within the broad frame-
work of constructional system theory. The polemically nominalistic stance of
the paper is declared in its notorious opening sentence (‘We do not believe in
abstract entities’ (Goodman and Quine 1947: 105)), and this declaration pro-
vided the philosophical foundation for the entire scheme. In order to motivate
their extreme stance, Goodman and Quine explain their refusal to tolerate such
notions as follows:
This statement indicates that, for Goodman and Quine (as for Leśniewski before
them) the main motivation for a nominalistic approach to constructional system
theory was the belief that, by rejecting abstract entities, the infamous paradoxes
that had disrupted all attempts to place mathematics on a secure (set-theoretical)
foundation might be avoided. Consequently, in this paper, Goodman and Quine
make a clear distinction between ‘platonistic logic’ (i.e., logic that admits
abstract entities such as sets) and ‘nominalistic logic’ (i.e., logic that does
not admit abstract entities such as sets), and the first sections of the paper are
devoted to outlining some of the strategies that must be employed in order to
convert platonistic statements into nominalistic statements. For instance, the
platonistic statement ‘Class A has three members’ assumes the existence of a
finite class (i.e., set), which is an abstract entity. However, if there are three
distinct objects x, y, and z, such that an object is in A if and only if it is x
86 Mathematical linguistics
∃x∃y∃z[x = y ∧ y = z ∧ x = z ∧ ∀w[w ∈ A
≡ w = x ∨ w = y ∨ w = z]] (3.16)
Clearly, (3.16) assumes the existence of nothing other than concrete individuals
in addition to the basic logical operators.
Having outlined their basic approach in this fashion, Goodman and Quine pro-
ceed to develop in detail the nominalistic syntax they require in order to remove
all ‘taint’ of platonism from their system (Goodman and Quine 1947: 107). The
syntax they create is inscriptional by design since it is ultimately concerned with
strings of physical marks rather than with abstract entities of any kind. Conse-
quently, as the foundation of the extralogical basis of their system, Goodman
and Quine define six so-called ‘shape-predicates’. These are predicates that
take concrete individuals as arguments and which specify the shape of a par-
ticular character. For instance, the predicate ‘Vee x’ means ‘the object x is a
vee (i.e., a v-shaped inscription)’ (Goodman and Quine 1947: 112). The other
characters defined in this way are: ’, (, ), |, and . In addition, a concatenation
relation, C, is assumed where ‘x C yz’ is understood to mean that ‘x and y
are composed of various characters of the language . . . and that the inscription
x consists of y followed by z’ (Goodman and Quine 1947: 112). The role of
concatenation is crucial in this system. As mentioned in section 3.5, Quine had
already explored the idea that the whole of arithmetic could be founded upon
the basis of concatenation alone, and, the inclusion of the C predicate in ‘SCN’
suggests that the notion of concatenation was considered to be fundamental also
to the formal syntax of a nominalistic object language. The final two predicates
used are introduced as ‘Part’ and ‘Bgr’, and are taken to mean ‘is part of’ and
‘is bigger than’ respectively. More precisely, ‘x Part y’ means ‘x, whether or
not it is identical with y, is contained entirely within y’, while ‘x Bgr y’ means
‘x is spatially bigger than y’ (Goodman and Quine 1947: 112–113).
Having presented the basic syntax of their object language, Goodman and
Quine then provide definitions of auxiliary predicates that are constructed in a
purely nominalistic manner from their initial predicate set. For instance, a four-
part predicate can be defined using the existential operator, logical conjunction,
and the concatenation operator:
Clearly the implication here is that natural languages and logical languages
differ considerably, at least as far as considerations of meaning are concerned,
and this view should be situated in the context of the general move towards
the further integration of natural and logical languages that has already been
discussed with reference to the work of Bloomfield, Ajdukiewicz, Harwood,
and Bar-Hillel. More specifically, White argues that issues of meaning are
fundamentally problematic in natural languages, though they can (usually)
be resolved in logical languages. As will be shown in sections 4.4 and 4.5,
Quine’s assault upon, and Goodman’s defence of, logical empiricism, along
with Quine and White’s questioning of the analytic–synthetic dualism, all stim-
ulated Chomsky while he was in the process of developing TGG, and it is crucial
to recognise that these developments were largely provoked by the constructive
nominalist research programme that Goodman and Quine had propounded in
the 1940s.
This passage suggests that, at this stage in his career, Carnap was convinced
that philosophy could be replaced by the systematic study of logical syntax.
However, in part 4 he broadens his basic approach and attempts to outline
the formal syntactic structure of language in general. It is crucial to note that
Carnap consistently views artificial languages as forming a well-defined subset
of natural languages, though he makes it clear that his intention is not to describe
the syntax of natural language.
The types of rules that Carnap refers to as ‘formation’ and ‘transformation’ rules
will be discussed at some length below, but it is necessary to stress that the above
passage reveals the domain of Carnap’s enquiry: instead of focusing on natural
languages, he concentrates on the task of defining and constructing formal
artificial languages, and, in order to accomplish this, he utilises assumptions
and techniques derived from (Hilbertian) Formalism. Despite this emphasis,
Carnap does not entirely preclude the possibility that the techniques he develops
in LSL may be of relevance for theories of natural language. Indeed, he states
explicitly that
The method of syntax which will be developed in the following pages will not
only prove useful in the logical analysis of scientific theories – it will also help
in the logical analysis of the world-languages. Although here, for the reasons
indicated above, we shall be dealing with symbolic languages, the syntactical
concepts and rules – not in detail but in their general character - may also be
applied to the analysis of the incredibly complicated world-languages. (Carnap
1937[1934]: 8)
As will be shown in chapters 4 and 5, this basic belief that ‘symbolic’ and
‘world’ languages could (in principle) be analysed using the same fundamental
Formal linguistic theory 91
techniques exerted a potent influence over certain linguists during the following
years.
It was suggested above that LSL reveals the influence of (Hilbertian) Formal-
ism, and Carnap’s indebtedness to Formalism is explicitly acknowledged several
times; indeed, the influence of Hilbert pervades the entire book. For instance,
Carnap specifically credits Hilbert with developing the theory of formal lan-
guages, observing that ‘[t]he point of view of the formal theory of languages
(known as ‘syntax’ in our terminology) was first developed for mathematics by
Hilbert’ (Carnap 1937[1934]: 1), and he utilises Hilbert’s notion of a formal
system extensively. For instance, at the very beginning of part 1 Carnap offers
the following definition of the term ‘formal’:
and, obviously, this quotation evinces the same concern for meaning-less sym-
bol manipulation as the various pedagogic texts considered in 2.8, with the
result that Carnap’s formal languages consist of meaning-less formulae derived
ultimately from primitive symbols by means of rules in a characteristically
Formalist manner; that is, the formal languages are defined solely in terms
of the syntactic structure of the sentences they produce, and the meanings of
the resulting formulae and primitive symbols are not considered. In order to
emphasise this point, Carnap considers the sentence ‘Pirots karulize elatically’
and states that this sentence can be parsed accurately as a Noun+Verb+Adverb
sequence even though the words are all unfamiliar (Carnap 1937[1934]: 2),
thus demonstrating that sentences can be exhaustively analysed solely in terms
of their formal syntactic structure even if the meaning of the individual words
is not known. This type of argument, which (according to Carnap) affirms the
separation of meaning and syntax, proved to be influential, and it will reappear
when TGG is discussed in section 5.7.
There is another aspect of the theory Carnap develops in LSL that is relevant to
the development of TGG. As mentioned above, Carnap distinguishes between
formation rules and transformation rules, and he introduces these terms in a
passage that considers the possibility of reducing logic to ‘syntax’:
Logic will become syntax, provided that the latter is conceived in a sufficiently
wide sense and formulated with exactitude. The difference between syntacti-
cal rules in the narrower sense and the logical rules of deduction is only the
92 Mathematical linguistics
G1 ∨ G2 (3.20)
G 2 , G 1 ⊃ G3 (3.21)
where ‘⊃’ is the symbol Carnap uses to indicate implication. This rule states
that sentence G3 can be obtained from the sentences G1 and G2 by means
of implication, although G3 does not constitute a simple combination of G1
and G2 . Consequently, (3.21) constitutes a transformation rule, and it should
be obvious that this rule is simply a restatement of Hilbert’s proof-theoretical
schema, given as (2.10) above.
Transformation rules are used extensively in LSL, and Carnap repeatedly
emphasises their centrality. The following passage is typical:
Once again, this passages stresses the fact that in Carnap’s framework trans-
formation rules are purely rules of logical inference that indicate a ‘logical
relation’ between particular sentences, with one sentence following as a con-
sequence of another sentence (or group of sentences). In case this summary
gives the impression that Carnap was devoid of any qualms concerning the
nature of logical inference, it is necessary to indicate that he was fully cog-
nizant of the many difficulties that beset his methodology. Indeed, Carnap was
happy to acknowledge that the notion of ‘consequence’ had to remain unde-
fined in his system, though the more restrictive notion of derivability could be
established.
later argues that, in considering the ‘characters’ (i.e., symbols) used in logical
and mathematical discourse, he has not ‘left the domain of language’ since
In general, to be sure, the separate characters have been agreed upon as sub-
stitutes for specific words or phrases. In many cases, however, we manage
best by ignoring the values and confining ourselves to the manipulation of
the written symbols; systems of symbolic logic, especially, may be viewed,
in a formal way, as systems of marks and conventions for the arrangement
of these marks . . . our formal systems serve merely as written or mechanical
mediations between utterances of language. (Bloomfield 1955[1939]: 262)
This passage, which could easily have come from one of the textbooks discussed
in 2.8, suggests that Bloomfield was essentially convinced of the validity of the
Formalist approach to mathematics. At the very least, it implies that Bloomfield
accepted the Formalist dictum that ‘we manage best’ (to use his own words) if
we focus on syntactic manipulation and ignore considerations of meaning. The
implications of this aspect of Bloomfield’s work are considerable and have never
been adequately discussed. Indeed, an exhaustive exploration of Bloomfield’s
appreciation of Formalism could well help to explain why so many young lin-
guists in the 1950s found that the techniques of structural linguistics appeared to
be compatible with the techniques employed by the formal sciences. In essence,
as the above passage demonstrates, Bloomfieldian linguistics and the formal
sciences were both shaped by Formalism during the 1930s, and the effects of this
influence are readily apparent in Bloomfield’s work. For instance, to consider
one example, it is well known that Bloomfield repeatedly expressed scepticism
concerning the role of meaning in linguistic theory. A standard expression of
this mistrust, taken from Language, runs as follows: ‘The statement of meaning
is . . . the weak point in language-study, and will remain so until human knowl-
edge advances very far beyond its present state’ (Bloomfield 1933: 140). In the
past, attempts to account for this scepticism have focused upon ideas concern-
ing syntax and semantics within linguistics, and upon the relationship between
linguistics and psychology.26 While there is no doubt that linguistics and psy-
chology were both responsible for determining the direction of Bloomfield’s
thought in many ways, it is certainly possible that some of his ideas concerning
the role of meaning in linguistic theory were directly influenced by his knowl-
edge of Formalism (and/or vice versa), which appeared to advocate the manip-
ulation of meaning-less symbols extracted from their semantic context. While
it would be needlessly excessive to claim that Bloomfield mistrusted linguistic
meaning primarily because he was intrigued by the methodology of Formal-
ism, it might well have been the case that his understanding of the foundational
96 Mathematical linguistics
The linguist naturally divides scientific activity into two phases: the scientist
performs ‘handling’ actions (observation, collecting of specimens, experi-
ment) and utters speech (report, classification, hypothesis, prediction). The
speech-forms which the scientist utters are peculiar both in their form and in
their effect upon hearers. (Bloomfield 1935: 499)
He later clarifies the nature of this peculiarity by observing that the language of
mathematics can only be understood after ‘severe supplementary training’, and
that utterances in such a language have the curious effect of causing the hearers
to ‘respond uniformally and in a predictable way’ (Bloomfield 1935: 499).
Clearly, therefore, the language of science differs significantly from natural
language, and the speech-forms of scientific language appear to constitute ‘a
highly specialized linguistic phenomenon’ (Bloomfield 1935: 500). It is at this
point that Bloomfield’s ambitious agenda starts to reveal itself. The following
passage is crucial:
To describe and evaluate this phenomenon is first and foremost a problem for
linguistics. The linguist may fail to go very far towards the solution of this
problem, especially if he lacks competence in the branches of science other
than his own. It is with the greatest diffidence that the present writer dares
to touch upon it. But it is the linguist and only the linguist who can take the
first steps towards its solution; to attack this problem without competence in
linguistics is to court disaster. The endless confusion of what is written about
the foundations of science or of mathematics is due very largely to the author’s
lack of linguistic information. (Bloomfield 1935: 500)
The central idea here is transparent: the complex and acrimonious arguments
that had come to characterise the mathematical foundations crisis debates in
the 1920s and 1930s could be resolved if only the participants were able to
view the problem from a linguistic perspective. Indeed, in Bloomfield’s words,
‘the linguist and only the linguist’ can intervene in order to resolve the dis-
putes. Obviously, this is a bold and startling claim, hence Bloomfield’s self-
confessed ‘diffidence’, but the proposal is serious nonetheless. Since (infuriat-
ingly) Bloomfield does not cite specific sources in his discussion, the precise
causes of his dissatisfaction with existing proposed solutions to the foundations
crisis can only be guessed. It should be recalled, though, that, as mentioned
previously, introductory texts such as Young’s Lectures on the Fundamental
Concepts of Algebra and Geometry predated the main foundational debates,
and consequently did not contain detailed discussions of the main disagree-
ments, suggesting that Bloomfield acquired his knowledge of these debates
from primary sources. As mentioned in section 3.5, some foundational issues
were addressed in certain works produced by the members of the Vienna Circle,
98 Mathematical linguistics
Before continuing with the footnote it is worth pausing to clarify the discus-
sion. As should be apparent, Grelling’s heterological paradox is closely related
to Russell’s paradox (discussed in section 3.4 above), the main difference being
that, rather than being situated in the context of set theory, Grelling’s paradox
enables Bloomfield to view the problem from the perspective of natural lan-
guage so that it can be assessed from a different standpoint. However, a mere
restatement of a known difficulty is one thing, while a specific proposal for its
resolution is quite another, yet, as the footnote continues, this is precisely what
Bloomfield attempts:
The fallacy is due to misuse of linguistic terms: the phrase ‘an adjective which
describes itself’ makes no sense in any usable terminology of linguistics; the
example of short illustrates a situation which could be described only in a
different discourse. E.g.: We may set up, without very rigid boundaries, as to
meaning, various classes of adjectives. An adjective which describes a pho-
netic feature of words is morphonymic (e.g., short, long, monosyllabic). A
Formal linguistic theory 99
Having made the attempt, the present writer has reached the conclusion that
such a study, apart from its linguistic interest, leads to the solution of cer-
tain problems that have baffled non-linguistic attack – the problems which
concern the foundations of mathematics. If this conclusion is justified, the
following pages should be of wider than linguistic interest. (Bloomfield
1970a[1937]: 335)
The members of the summation series can be obtained one by one, but we
have no finite formula for the direct naming or recognition of these members.
To prescribe the naming, in this form, of an irrational number, is to insist that
our hearers complete the recitation of an infinite class of speech-forms. This
fallacy is still current among mathematicians; we shall return to it in Chapter
22. (Bloomfield 1970a[1937]: 337)
ten digits are .5471111117. This formula for naming N1 , is stated in terms of
and its well-ordering: a digit of N1 can be named only if one first names
k digits of the kth R. Hence to calculate and recite digits of N1 to the end of
one’s patience is not to name a number: it is only the formula N1 , interpreted
as above, which names a number. (Bloomfield 1970a[1937]: 338)
Although this remnant of a larger discussion is opaque in places, the basic thrust
of the passage is clear: the act of enumerating the members of an infinite class
(i.e., set) is not the same as naming the set, and, presumably, in the remaining
chapters of the MS, Bloomfield sought to demonstrate that the paradoxes of set
theory could be obviated if this kind of linguistic distinction were systematically
observed.
When the remaining MS fragments were collected by Hockett in 1970 for
inclusion in A Leonard Bloomfield Anthology (which he was then editing), he
commented concerning the destruction of the manuscript:
I cannot refrain from expressing my regret at this loss. Had he [i.e., Bloom-
field] lived to rework the topic, benefiting from Professor Curry’s suggestions
(even if not accepting them all), some of his successors, who have concerned
themselves with the inter-relations of language and mathematics, might have
been helped to avoid various stupid errors. (Bloomfield 1970b: 334)
Unfortunately, Hockett does not name the linguists who had been guilty of
making ‘stupid errors’, nor does he indicate which particular mistakes he has
in mind. There can be little doubt, though, that some of the names that fea-
ture prominently in the remaining chapters of this book may well have been
Hockett’s intended targets.
Bloomfield’s ideas concerning the relationship between mathematics and lin-
guistics and, in particular, his Formalist tendencies (whether overtly or covertly
expressed) which emphasised the primacy of syntactic (rather than seman-
tic) considerations, exerted a profound influence over a whole generation of
linguists that came to maturity in the 1940s and 1950s. It is important to recog-
nise, though, that this desire for a more mathematical approach to linguistic
analysis was not solely confined to North America. For instance, to consider
just one European example, Bloomfield’s interest in recent advances in math-
ematics was shared by the heterogeneous group of researchers that constituted
the Copenhagen Circle, and especially by Louis Hjelmslev (1899–1965). In his
1943 book Prolegomena to a Theory of Language, for example, Hjelmslev cited
Bloomfield’s 1926 paper in a footnote when he refers to ‘transcendent kinds
of linguists’ who have attempted to construct ‘systems of axioms’ (Hjelmslev
1961[1943]: 6). Hjelmslev not only shared Bloomfield’s interest in axioms and
102 Mathematical linguistics
This passage indicates that, for Hjelmslev, there was a clear difference between
the concerns of metamathematics and linguistics. The two disciplines were asso-
ciated by their status as sub-branches of semiotics, but linguistics could never
be subsumed by metamathematics because semantic considerations dominate
in the former while they are non-existent in the latter. Nevertheless, the above
passage clearly indicates that Hilbertian Formalism passed from the realm of
mathematics into the realm of linguistics in Europe as well as North America,
and that, for Hjelmslev (at least), the transmission of ideas was mediated by
‘the Polish logicians’ and Carnap.
While certain European linguists, such as Hjelmslev, were gradually devel-
oping a more formal approach to linguistic theory during the 1930s and 1940s,
similar developments were occurring in North America, and since TGG is part
of the story of North American linguistics, the focus of this discussion must
once again fall upon the United States. As already indicated, the influence of
Formalism is perhaps most apparent in the work of the post-Bloomfieldians
– a disparate group of linguists that included Bloch, Hockett, Yuen Ren Chao
(1892–1982), Rulon Wells (b. 1919), Martin Joos (1902–1978) and, of course,
Harris32 - who were, in many different ways, directing their research towards the
Formal linguistic theory 103
In view of the fact that methods as mathematical as the one proposed here have
not yet become accepted in linguistics, some apology is due for introducing
the procedure. However, the advantage which may be gained in explicitness,
and in comparability of morphologies, may offset the trouble of manipulating
the symbols of this procedure. (Harris 1946: 161.n1)
The phrase ‘manipulating the symbols’ suggests the Formalist tendency of Har-
ris’ thinking (i.e., morphemic analysis is essentially viewed as an exploration
of symbol permutations), and Harris’ apology for using this kind of ‘mathemat-
ical’ procedure suggests that he himself believed that the nature of linguistic
research had begun to change. In the same way, many of the post-Bloomfieldian
papers written during the 1940s and 1950s reveal fragments of their intellectual
heritage which serve to illuminate the various influences that were prompting
these changes. However, it was only with hindsight that Harris himself compre-
hensively acknowledged the multifarious influences that had inspired his own
research during this period:
Post and Kurt Gödel, in recursive function theory, and from a somewhat dif-
ferent direction in the Turing machine and automata theory . . . In linguistics,
the ‘distributional’ (combinatorial) methods of Edward Sapir and Leonard
Bloomfield were hospitable to this approach. Cf. also Nelson Goodman, The
Structure of Appearance . . . (Harris 1991: 145)
This catalogue of influences indicates that Harris was well aware of the major
developments in the philosophy of mathematics that had dominated the first
decades of the twentieth century, and that he believed these ideas had directly
influenced his own approach to the study of language. Also, once again we
encounter the claim that techniques derived from the formal sciences were
considered to be compatible with the distributional methods employed by the
post-Bloomfieldians. For Harris, the ‘distributional methods’ used in linguistics,
and originally associated with Bloomfield and Sapir, seemed to be ‘hospitable’
to the techniques of logical analysis, and this recalls Bar-Hillel’s assertion that
Ajdukiewicz’s system of logical syntax could be combined with mechanical dis-
covery procedures; and, as suggested above, Bloomfield’s Formalist tendencies
may well have been responsible for preparing the ground for this compatibil-
ity. In addition, it is worth noting that, in the same paragraph as that partly
quoted above, Harris goes on to refer explicitly to two contemporaneous text-
books, Kleene’s Introduction to Metamathematics and Church’s Introduction
to Mathematical Logic, citing them as sources that were used by linguists at
the time. Once again, therefore, the influence of these pedagogic texts upon
the linguistics community is apparent. Indeed, the relationship between logi-
cians and linguists was growing closer during the 1940s, and, in the context of
this rapprochement, it is significant that the manuscript of Church’s book was
proofread by the philosopher-cum-linguist Rulon Wells (Church 1944: vi).
Since a consideration of Carnap’s LSL has already occupied part of this
section, and since the text was to influence the development of TGG directly,
it is worth emphasising the fact that, although Carnap is not one of the authors
Harris cited explicitly in his 1991 reflections, several publications from the
1950s reveal the extent of Harris’ knowledge of LSL. For instance, in MSL,
Harris considers the general problem of linguistic analysis and observes in a
footnote:
parts of the actual speech occurrences as their elements, they set up very simple
elements which are merely associated with features of speech occurrences.
(Harris 1951: 16.n17)
Obviously, this passage indicates that Harris was familiar with Carnap’s work.
However, more than this, the content of the passage indicates that, far from
wholeheartedly advocating the adoption of techniques from logic for the pur-
poses of analysing natural languages, Harris was keen to stress the differences
that distinguish linguistics and logic. Essentially, the main difference appears
to be that linguists deal with natural languages, while logicians create artificial
languages. Nevertheless, this does not imply that, for Harris, logical languages
and natural languages were entirely different kinds of systems, rather that differ-
ent types of researchers (i.e., linguists and logicians) focus on different aspects
of the problem.
As suggested above, formal syntactic theories (especially the theory outlined
by Carnap in his LSL) exerted an influence on Bloomfield, Hjelmslev, Wells,
Harris, and many other linguists, particularly during the 1940s and 1950s. How-
ever, perhaps the most enthusiastic adherent of Carnapian syntactic theory in
linguistic circles was the irrepressible Bar-Hillel. Bar-Hillel’s undergraduate
training had been in mathematics and philosophy (not an uncommon route to
the theoretical study of language in the days when there were few undergraduate
linguistics courses on offer) and, in his preface to Language and Information
(1964), he provided a detailed overview of his own intellectual development.
Once again there is a polymathic cascade of names – Quine, Tarski, Bloomfield,
Reichenbach, Ajdukiewicz, and so on – but, in particular, he recalls the influ-
ence of Carnap. He first encountered LSL during the academic year 1936/1937
and,
For the next couple of years, I was seldom seen without a copy of this book
under my arm. My fellow students dubbed it ‘Bar-Hillel’s Bible’. It was doubt-
less the most influential book I read in my life, and a good part of my work is
directly or indirectly related to it. (Bar-Hillel 1964: 2)
Clearly, then, Bar-Hillel’s peers considered his devotion to LSL to have been
equivalent to a religious conviction, and some of the main tenets of Bar-Hillel’s
logico-syntactic faith are discussed in relation to TGG in section 4.5. It is of
interest, though, that Bar-Hillel had also encountered Bloomfield’s 1939 essay
concerning the scientific status of linguistics, and that this text too influenced the
development of his thinking. Once again his own comments are illuminating:
106 Mathematical linguistics
I think that the only work by a modern professional linguist I had studied
in some depth before these talks [i.e., talks with Harris in the early 1950s]
was Bloomfield’s little contribution to the Encyclopedia of Unified Science,
published in 1939. This booklet showed a surprising convergence between
ways of thinking of at least certain circles of American linguists and those of
say, Carnap, and I made a mental note to pursue this issue further sometime.
But only in 1951 did I find the time to do so. (Bar-Hillel 1964: 4)
This passage indicates that, for Bar-Hillel at least, Bloomfield’s work suggested
the possibility of integrating techniques from Carnapian logical syntax with the
methodology of linguistics. As already indicated, this kind of project was cer-
tainly not unique since, by the early 1950s, there was a general tendency in
certain linguistic circles to seek formal linguistic theories (particularly syntac-
tic theories) which advocated meaning-less symbol manipulation rather than
contentual, semantics-based analysis, and, as discussed above, Carnap’s LSL
was a prominent text that appeared to provide formal analytic techniques that
could be usefully transferred from the realm of logical syntax into the sphere
of linguistics.
108
Biography and influences 109
mind, the basic chapter plan can be summarised as follows. In section 4.2 some
biographical information about Chomsky is presented, with particular attention
given to his contact with some of the individuals mentioned in chapter 3. This
enables direct connections between his work and that of his most influential pre-
decessors to be established. In section 4.3 the influence of Goodman is explored
by examining the way in which Chomsky reformulated the notion of simplicity
in syntactic terms. This consideration of simplicity naturally invites an assess-
ment of other aspects of Goodman’s influence upon the young Chomsky, and
section 4.4 explores his brief flirtation with constructive nominalism. Finally,
in section 4.5 some of the arguments concerning the relationship between syn-
tactic theory and the formal sciences are considered. In particular, the debate
between Bar-Hillel and Chomsky concerning the relationship between logic
and linguistic theory that took place during the years 1954–1955 is assessed,
since it provides valuable insights into Chomsky’s whole approach to syntactic
theory.
This perception of Harris’ work seems to have been commonplace. For exam-
ple, in the 1960s Hockett had referred to Harris’ ‘theoretical nihilism’ (Hockett
1968: 35). However, despite Harris’ early interest in Chomsky, the two men
Biography and influences 111
even at this early stage [i.e., early 1950s], Chomsky was producing highly
original work, which diverged fundamentally from Harris’. In his B.A. thesis
he was doing things that were, in his own words, ‘radically at odds with
everything in structural linguistics . . . which is why [both the thesis and The
Logical Structure of Linguistic Theory] were published only 30 years later’.
The thesis was ‘as different from structural linguistics as anything could be’,
which was why ‘Harris never looked at it and no one in the field reacted to it’.
(Barsky 1997: 53)
In a later passage Chomsky suggests that Harris considered the nascent theory
of TGG to be ‘crazy’ (Barsky 1997: 54), although no evidence to support this
claim is provided. Strangely, the basic pattern of Chomsky’s relationship with
Harris (or at least Chomsky’s own account of it) is identical to the basic pattern
of his relationship with Goodman. At first Goodman encouraged and supported
Chomsky, helping him to obtain a Junior Fellowship at Harvard, but then the
association turned sour. The following passage is also taken from Barsky’s
account, mainly because (once again) it weaves together some of Chomsky’s
own words. After mentioning (with no details) that Chomsky began to study the
work of Carnap, Frege, Whitehead, and Wittgenstein (no mention of Russell?),
Barsky continues,
the readings Chomsky now undertook gave him a fresh perspective that his
teacher Nelson Goodman considered to be ‘completely mad’. When Goodman
found out about Chomsky’s work in the mid-1960s, he apparently ended their
friendship, even though, as Chomsky says, they’d ‘been quite good friends
until he learned about this, which he regarded somehow as a personal betrayal’.
(Barsky 1997: 54)
There are various problems with this account. First, it is not at all clear why
Goodman would have been offended by any ideas Chomsky might have gleaned
from Carnap, Whitehead, and Wittgenstein (or even Russell). Second, as will be
shown in section 5.3, if anything caused Chomsky to find a ‘fresh perspective’
during the early 1950s, it was Goodman’s own work. Third, it is not clear
why it took Goodman ten years to find out about Chomsky’s research into
syntactic theory. By the mid-1960s Chomsky’s name was already well known
and Goodman must have been living as a hermit not to have heard something
about TGG before c. 1965. Curiously, though, Chomsky’s memories of his
friendship with Quine outline a similar pattern: at first Quine had encouraged
112 Systems of syntax: 1951–1955
him (for instance, by reading his 1953 paper before it was submitted), but then
‘lost interest’ (quoted in Barsky 1997: 93) in Chomsky’s work.
This general pattern of initial closeness followed by sudden separation is
clearly of interest and it raises various questions: why did Chomsky come to be
spurned by his early mentors, and precisely which aspects of his work offended
them? Whatever the truth of Chomsky’s recollections, though, as will be shown
in the next two chapters, there can be no doubt that, during the early 1950s,
Chomsky was working closely with Harris, Goodman, Quine, and others, and
that he was directly influenced by some of their ideas. Accordingly, the task of
revealing traces of these influences in his work from the 1950s is accomplished
in the following sections, and it may be the case that an exploration of this
kind proves to be of greater interest than a protracted attempt at retrospective
psychological analysis.
linguistic units and, in particular, must distinguish and characterise just those
utterances which are considered “grammatical” or “possible” by the informant’
(Chomsky 1979b[1951]: 1). Second, the grammar must either meet criteria that
are imposed by the ‘special purpose’ for which the grammar was created (e.g.,
pedagogic utility), or else, if there is no such purpose, it must meet require-
ments of ‘simplicity, economy, compactness, etc.’ (Chomsky 1979b[1951]: 1).
From the very beginning, then, Chomsky seems to have been persuaded that
considerations of simplicity were intimately involved in the processes of gram-
mar construction, and the task of unearthing the sources of these ideas is fairly
simple, mainly because Chomsky never attempted to conceal his intellectual
debts. For instance, in MMH he explicitly cites Goodman’s 1943 paper, ‘On the
Simplicity of Ideas’, and paraphrases Goodman’s argument against simplicity
being considered merely an aesthetic luxury in a constructional system (quoted
in full in section 3.5 above). With particular reference to syntactic theory, one
of the implications of Goodman’s views concerning the critical importance of
simplicity criteria in constructional systems is that the reasons for wanting a
grammar to be as simple as possible are the same as for wanting a grammar
at all. Crucially, though, it should be noted that this focus upon theory-internal
economy considerations was not encouraged by Harris, who was the other main
influence upon the form and content of MMH. For instance, in MSL Harris had
clearly stated that ‘it is a matter of other than descriptive purpose how compact
and convenient the formulation is, or what other qualities it may have’ (Harris
1951: 9), and he had gone on to claim that
It therefore does not matter for basic descriptive methods whether the system
for a particular language is so devised as to have the least number of elements
(e.g., phonemes) or the least number of statements about them, or the greatest
over-all compactness, etc. These different formulations differ not linguistically
but logically. They differ not in validity but in their usefulness for one purpose
or another (e.g., for teaching the language, for describing its structure, for
comparing it with genetically related languages). (Harris 1951: 9.n8)
For Harris, then, practical utility, rather than considerations such as simplicity,
was the guiding principle of grammar construction, and this may well be a good
example of Harris’ aforementioned ‘theoretical nihilism’ (Hockett 1968: 35).
Whether this is the case or not, the clear implication of the above passage is
that Harris deemed issues such as grammatical ‘compactness’ to be extraneous
to the task of linguistic analysis, or, at the very least, he was convinced that
such considerations should only be assessed with reference to the utility of the
grammar. In other words, if simplifications facilitate the use of the grammar as
114 Systems of syntax: 1951–1955
a pedagogic tool, then they are valuable; if not, then they are of no real interest.
The clear juxtaposition of Goodman’s repeated emphasis on the importance of
theory-internal economy and Harris’ apparent indifference to the same must
have intrigued Chomsky at the time, and, indeed, MMH can be viewed partly
as an attempt to introduce Goodman-style simplicity criteria into a Harris-
style analytic framework. Already, therefore, Chomsky’s remarkable ability to
reconcile and synthesise existing techniques is clearly apparent in the perceived
analogy between syntactic analysis and constructional system theory.
Having outlined the basic problem in this general fashion, Chomsky goes on
to distinguish between ‘discovery’ processes (e.g., Harris’ distributional proce-
dures) and processes of ‘description’. The distinction between these processes
is clear: discovery processes can be used to determine the set of grammatical
sentences in a corpus, while processes of description can be used to analyse the
form of those utterances. In this context, it is worth recalling that Bar-Hillel’s
revival of Ajdukiewicz’s system of syntactic analysis (discussed in section 3.4)
was intended to provide a method of ‘description’ (Bar-Hillel 1953b: 47) and
was not intended to constitute a discovery procedure. In the early 1950s, then,
the distinction that Chomsky’s draws was certainly not unique. However, he pro-
ceeds to extend the argument by suggesting that, although this distinction might
be easy to make when a grammar is constructed for a ‘special purpose’, it is less
useful when it is created without a particular use in mind, for, in this case, there
are no arbitrary constraints imposed upon the grammar. Consequently, in the
latter case, the grammar must be constructed ‘solely in accordance with consid-
erations of elegance’ (Chomsky 1979b[1951]: 2). Indeed, Chomsky goes on to
suggest that the very notion of ‘elegance’ itself threatens to undermine the basic
distinction between discovery and description processes, since such considera-
tions are fundamental to the task of discovery and to the task of description. In
the light of this observation, it is worth emphasising that, from the very begin-
ning, Chomsky’s focus was upon syntactic analysis for the sake of syntactic
analysis. He appears never to have been especially interested in the practical
(i.e., pedagogical) applications of the sort that so preoccupied Harris.
Despite the fact that Chomsky explicitly places the issue of grammatical sim-
plicity at the core of the theoretical framework he outlines in MMH, he does not
provide an exhaustive discussion of the topic. He observes that merely counting
the number of rules in the grammar is an inadequate way of measuring sim-
plicity, and he refers somewhat obliquely to an unpublished paper of his own
that considers the topic, but he does not summarise the contents of that paper in
his thesis.3 Nevertheless, Chomsky maintained his interest in simplicity mea-
sures long after his M.A. thesis was submitted, and the most comprehensive
Simplicity and grammar 115
discussion of the subject was offered in chapter 4 of his manuscript The Logical
Structure of Linguistic Theory (hereafter LSLT), the first draft of which was
completed by early 1955.4 In this work the importance of grammatical simplic-
ity is stressed repeatedly and (once again) it is considered to play a fundamental
role in the task of grammar construction.
This passage neatly summarises the core motivation for an assessment of sim-
plicity in the context of linguistic theory: simpler grammars are preferred
because they capture more general linguistic patterns than more complex gram-
mars. Clearly, this assumption echoes Goodman’s pro-simplicity arguments,
which were discussed in detail in section 3.5. For instance, Goodman had
asserted that ‘the more the set of postulates can be reduced without becom-
ing inadequate, the more comprehensively will the system exhibit the network
of interconnections that comprises its subject matter’ (Goodman 1943: 107),
and this observation is clearly related to various passages in LSLT such as that
quoted above. Specifically, in linguistic terms, a measure of grammatical sim-
plicity is required in order to permit the ‘best’ grammar for a given corpus to
be selected from among all other possible grammars, and the task of creating
such a simplicity measure is precisely the one that Chomsky sets for himself
in chapter 4 of LSLT. Before discussing his treatment of this topic in detail,
though, it is worth highlighting his belief (already expressed in MMH) that the
technical issue of grammatical simplicity is not merely a superficial adjunct to
linguistic theory. Indeed, Chomsky bluntly asserts that the notion of simplicity
is of fundamental importance, and he admits that his thinking concerning this
issue has been guided by the work of both Goodman and Quine.
presented in LSLT in some detail, since it reveals the manner in which Chomsky
modified some of Goodman’s and Quine’s ideas so that they could be used to
facilitate the specific task of syntactic analysis.
At the outset of his discussion of grammatical simplicity in chapter 4 of LSLT
Chomsky openly confesses that his current thinking on this topic is still at a
preliminary stage and that, therefore, it is ‘sketchy and incomplete’ (Chomsky
1975[1955]: 116). Despite this, he is sufficiently confident about the issues
involved to be able to propose a series of consolidation rules that converts a
given grammar into a maximally condensed form. Once again, this emphasis on
rule-driven processes recalls Harris’ interest in automatic discovery procedures
(mentioned in section 3.7), as well as Bar-Hillel’s enthusiasm for syntactic
categories that enable a sentence to be analysed ‘mechanically’ (discussed in
section 3.4). In the present context, then, Chomsky’s desire for an algorithmic
approach to simplicity measures is in keeping with the general drift of syntactic
research in the 1950s. In addition, though, the search for a mechanical procedure
that would permit an automatic measurement of grammatical simplicity again
echoes Goodman who, in his 1943 paper, advocated the avoidance of intuition
when determining the respective simplicity of various bases, favouring instead
an ‘automatic procedure’ (Goodman 1943: 108). Given the fact (discussed at
length in section 5.4) that the definition of a linguistic level offered in LSLT
draws heavily upon constructional system theory, it would be reasonable to sus-
pect that, following Goodman, Chomsky would focus on the task of reducing
the number of primitive relations associated with a particular syntactic level.
However, since each level of the general system outlined in LSLT uses con-
catenation as its sole primitive relation, reducing the number of such relations
was not a feasible option. Consequently, Chomsky realised that, if a measure of
grammatical simplicity were ever to function as a valid evaluation procedure, it
must operate upon something other than the extralogical bases of the linguistic
levels.
Accordingly, developing the scheme outlined in MMH, Chomsky argued that
the basic formal components of the grammatical theory developed in LSLT that
must be involved in any technical definition of simplicity are the rewrite rules,
called ‘conversions’ in the early TGG literature. In the standard TGG notation,
rewrite rules take the form ‘X → Y ’, and they are used to permit the analysis
of larger syntactic units into smaller sub-components. For instance, a standard
P-level (i.e., phrase-level) conversion is ‘Sentence → NP VP’, which states
(obviously) that a sentence can be rewritten as a noun phrase followed by a
verb phrase. It is reasonable that such rules should be involved in the techni-
cal definition of a simplicity measure, since a measure of any kind necessarily
118 Systems of syntax: 1951–1955
adeg (4.1)
bdg (4.2)
cd f g (4.3)
It is clear that (4.4) is more general than the separate strings in (4.1)–(4.3),
and that it is not simply a vacuous tautological restatement of pre-existing
facts, since it makes more particular claims about the distributional properties
of the formal language used in the example. As Chomsky observes, since this
kind of consolidation can be applied to both sides of a given conversion rule,
these notational manipulations enable a maximally consolidated grammar to be
constructed in such a manner that real complexity reduction is assured.
There is a great advantage in giving the principle of ordering once and for all
in the general theory. Otherwise, the gain in economy resulting from ordering
of rules in a particular grammar will be much reduced because of the need
to specify, within that grammar, the order of application of rules. The most
favourable situation is one in which we can linearly order the conversions in
such a way that all derivations can be formed by running through this sequence
from beginning to end. (Chomsky 1975[1955]: 125)
The implication here is that the conversions that constitute the grammar should
be so arranged that all grammatical utterances in a given corpus can be cor-
rectly derived simply by applying the various conversions in sequence. This
observation leads Chomsky to propose certain ‘optimality conditions’, which
he introduces as follows:
Putting it roughly, a grammar will meet these conditions if, when the rules are
given in a maximally condensed form, it is possible to arrange the resulting
statements in a sequence in such a way that:
(i) we can form all derivations by running through the sequence of rules from
beginning to end;
(ii) no conversion X → Y will appear twice in the sequence (i.e., no rule need
be repeated in several forms at various places in the grammar);
(iii) each conditioning context is developed to exactly the extent relevant for
the application of the rule in which it appears. (Chomsky 1975[1955]: 125)
grammars and transformational grammars was that the latter were able to deal
with unbounded dependencies more successfully than the former, while being
able to generate complex sentences, such as active–passive constructions, more
economically, and considerations such as these emphasised the centrality of
simplicity criteria in the earliest TGG research.
in the passage quoted above, since the ‘calculus of individuals’ mentioned there
is that developed by Goodman and Leonard during the mid-1930s, and Chom-
sky makes direct use of this system throughout his own paper, employing the
particular version of the calculus that appeared in Goodman’s SA. The implica-
tions of using this type of constructional system for the purposes of syntactic
analysis are many. In particular, the ‘inscriptional’ emphasis determines that
the sentences in a language must be manipulated as finite strings of symbols,
as in Formalist mathematics. Also, the rejection of such abstract entities as sets
ensures that all larger structures must be exhaustively analysed in terms of their
constituent elements.
Given the above, and given the freely confessed nominalist agenda of ‘SSA’,
Chomsky’s claim that a constructional approach to syntactic analysis could
help to resolve certain fundamental problems associated with the phenomenon
of homonymity is intriguing, if somewhat premature, since (as he himself
acknowledges) such problems do not manifest themselves in the simple for-
mal languages discussed in the paper, and (so far) he has never explored these
issues in subsequent work.9 Indeed, it is important to emphasise that through-
out ‘SSA’ Chomsky is solely concerned with formal languages, and openly
declares that ‘the present system as given here is not adequate for the analysis
of natural languages’ (Chomsky 1953: 243). The particular language he uses in
order to illustrate his theoretical points consists merely of ‘utterances’ such as
‘ab’ and ‘axd’, where each alphabetic character represents a morpheme in the
language. Obviously, this focus on such simple languages is in marked contrast
to Chomsky’s later work, which was predominantly concerned with problems
of natural language analysis; an issue that is further discussed in section 4.5.
As indicated above, in the details of its structure, the type of nominalistic con-
structional system that Chomsky develops in sections 2 and 3 of ‘SSA’ is merely
an implementation of the system presented by Goodman in SA, with the majority
of the definitions, axioms, and theorems presented being borrowed unaltered.
The paper’s main contribution, in fact, is simply the attempt to apply construc-
tional system theory to the task of syntactic analysis. Accordingly, following
Goodman, the core of the system outlined in the paper contains a logical basis
(i.e., standard symbolic logic of the PM variety) and an extralogical basis con-
sisting of primitive elements (i.e., inscriptions which are the basic morphemes)
and five primitive relations. It should be remembered that (as mentioned above)
the primitive elements are all considered to be ‘experimentally defined’, thus
revealing the strong empirical nature of the proposed analytic method.10 The
primitive relations include such predicates as ‘O’, where ‘a O b’ means a ‘over-
laps’ b (i.e., there exists an inscription that is part of both a and b), and these
124 Systems of syntax: 1951–1955
relations are all taken, unmodified, from SA, as Chomsky readily admits. He
then proceeds to construct a set of ‘elementary notions’ such as ‘SEG’, where
‘a SEG b’ means ‘a is a segment of b’, and is defined in terms of the logical
and extralogical bases as
a SEG b = ∀x[x O a → x O b] (4.5)
which indicates that a is a segment of b if, for all x, x overlaps both a and
b. Once again these definitions are Goodman’s (indeed, (4.5) is identical with
Goodman’s definition of his ‘Part’ relation, given as (3.15) above), and they
enable further axioms and theorems to be constructed. Finally, in section 3 the
manner in which Goodman’s constructional system can be utilised for the task
of syntactic analysis is outlined. The basic methodology is to construct ‘an
indefinite series of similarity relations’ of the form ‘Sn ’ (Chomsky 1953: 249),
where the subscript n ranges over numerals and indicates the length of the
morpheme sequence. These similarity relations can be used to group the mor-
phemes encountered in a given corpus into syntactic classes. To illustrate how
the process works, Chomsky uses a corpus of six sentences, ab, cb, de, f e,
axd, and cy f , and he shows that the morphemes a and c would be related by
the similarity relation S1 , as would the morphemes d and f , while a..d and
c.. f would be related by S2 , with x and y being related by S3 .
As the above summary indicates, the general methodology offered in ‘SSA’
suggests that the nominalistic constructional system presented by Goodman in
SA could be used in the context of syntactic theory as a discovery procedure
that would reduce the task of syntactic class assignment to an automatic pro-
cess, thus illustrating the potential correspondence between Harris’ discovery
procedures and constructional system theory. As mentioned in the introduc-
tion, though, Chomsky’s 1953 paper has generally been neglected by linguistic
historiographers, and, given the above summary, it is not difficult to see why
it has been forgotten (or perhaps intentionally ignored?). When viewed in a
cursory manner, the paper appears to be only tenuously connected to the type
of syntactic theory Chomsky was soon to develop in LSLT and which would
eventually become known as TGG. For a start, as mentioned above, ‘SAA’ is
solely concerned with simple formal languages, and, as shown in the next sec-
tion, Chomsky was soon to question the validity of arguments involving such
languages when used in the context of linguistic theory (as opposed to pure log-
ical research). Also, drawing as it does upon Goodman’s work in constructional
system theory, ‘SA’ appears to have a strong empiricist bias (though Hiorth, for
one, has doubted this (Hiorth 1974: 35)), since it focuses upon the automatic
assignment of morphemes in a specific corpus using distributional information.
Logic and linguistic theory 125
leading linguistics journal of the time for, as chapter 3 and the earlier sections of
this chapter have demonstrated, during the 1940s and 1950s the precise nature
of the relationship between logic and natural language was being reassessed
by various logicians and linguists, and certainly Bar-Hillel and Chomsky were
both active participants in this process of reassessment. Their exchange in the
mid-1950s began with the publication of Bar-Hillel’s paper ‘Logical Syntax
and Semantics’ (1954) in Language and continued when Chomsky published
‘Logical Syntax and Semantics: Their Linguistic Relevance’ (1955) (hereafter
‘LSS’) in the same journal. While considering the various arguments put for-
ward by Bar-Hillel, it is essential to remember that, as discussed in section 3.4,
he had recently revived the work of Ajdukiewicz, and had suggested that tech-
niques adapted from logical syntax and logical semantics could be used to
analyse natural language, thus laying the foundations for the function-based
syntactic theory now known as Categorial Grammar. Clearly, therefore, Bar-
Hillel was fully persuaded that logic could provide valuable insights into the
nature of linguistic structure and, though he later confessed that these ideas
were ‘deplorably naı̈ve’ (Bar-Hillel 1964: 3), they certainly inspired his 1954
paper, which must now be considered at length.
In general terms, Bar-Hillel’s paper constitutes a ‘plea for the reintroduction
of semantics into the theatre of operations of descriptive linguistics’ (Bar-Hillel
1954: 235–236), and his primary motivation seems to have been a profound dis-
satisfaction with the type of analytical methodology advocated by Harris in his
MSL. Revealingly, in order to articulate his frustration with distributional proce-
dures more precisely, Bar-Hillel quotes a lengthy passage from the 1937 English
translation of Carnap’s LSL, a book with which he had become obsessed.11 The
quoted passage is taken from the general introduction in which Carnap consid-
ers the relationship between syntax and logic, and in which he distinguishes
between formation and transformation rules. Carnap observes that ‘the prevalent
opinion’ (i.e., in 1934) was that syntax and logic are ‘fundamentally theories of
a very different type’, yet he goes on to challenge this presupposition by assert-
ing that both disciplines are ‘equally concerned with the relation of meaning
between sentences’, which in turn leads him to prophesy that ‘logic will become
part of syntax’ (Carnap 1937[1934]: 1–2). By means of this conviction, Car-
nap is able to speak coherently of ‘logical syntax’, a hybrid term that gestures
towards the interrelation of the two disciplines.
Having presented Carnap’s views in this direct manner, Bar-Hillel then con-
siders the way in which these ideas had been (and could be) developed by
linguists. He emphasises the fact that formation rules (in the Carnapian sense)
had been explored in some detail, citing Harris’ MSL and Fries’ The Structure
Logic and linguistic theory 127
and Bar-Hillel interpreted this to mean that linguistic units associated with
different distributions inevitably have different meanings. Consequently, he
concludes that, for Harris, units with identical distributions are considered to
be synonymous. This is admittedly a rather strong interpretation of Harris’
statement, but the inference is not entirely unreasonable. However, Bar-Hillel
rejects the idea that synonymy can be fully analysed in terms of distribution,
128 Systems of syntax: 1951–1955
arguing that, while a distributional analysis might be able show that (i) ‘oculist’,
(ii) ‘eye-doctor’, and (iii) ‘dentist’ can all appear in the same syntactic positions,
it can never indicate that (i) and (ii) have identical meanings, while this is
not the case either for (i) and (iii) or (ii) and (iii). This failure, according to
Bar-Hillel, is caused by the neglect of the truth-conditions that would provide
information about the meaning of the words. Consequently, since distributional
procedures ignore such semantic considerations, they provide an incomplete
basis for linguistic analysis.
Bar-Hillel’s conviction that semantic considerations can be incorporated into
syntactic analyses without the latter succumbing to ‘an infestation by meaning’
(Bar-Hillel 1954: 234) clearly has its roots in his well-attested interest in the
work of Carnap and the Lvov-Warsaw school of logicians. As discussed in sec-
tion 3.7, Carnap’s research into logical syntax had been influenced by Hilbert’s
proof theory and consequently, he had defined a number of formal languages
that would enable him to create logical systems that were free from semantic
considerations. However, in turn (and rather ironically), this work caused Car-
nap to reflect upon the relationship between formal linguistic systems of the
type presented in the LSL and the external world. In other words, his interest in
formal syntactic systems that attempted to avoid semantic considerations incul-
cated a profound interest in semantics, and for the next twenty years Carnap
focused his research upon this broad topic. In 1942 he published his Intro-
duction to Semantics while, in the following year, his Formalization of Logic
appeared, and both books explored various aspects of semantics in relation
to formal systems. The main emphasis in both texts is upon truth-conditional
interpretations. In other words, given a statement in a formal language, in order
to explore its meaning, Carnap argued that it was first necessary to know how
the universe would have to be configured in order for the statement to be true.
In Carnap’s treatment, the rules for semantic interpretation constitute a separate
and isolated part of the full linguistic system he develops: the syntactic com-
ponent outputs sentences in a formal language that are subsequently subjected
to semantic interpretation. In order to clarify his position, Carnap offers the
following definition of a semantic system:
The basic idea behind this schema is that any sentence will be classified as an
‘equivalence’ of the form (T ) if the letter ‘ p’ above can be replaced by a sen-
tence, and the letter ‘X ’ by the name of the sentence. Once again, the Hilbertian
notion of a metalanguage is dominant here since there is a clear distinction
between metalanguage and object language. The most famous example of a
sentence-name pair that is equivalent to (T ) is the sentence ‘Snow is white if
and only if snow is white’, and it is the concept of a metalanguage alone that
rescues this statement from tautology: ‘X ’ (i.e., ‘snow is white’) is true if and
only if it is actually the case that p holds (i.e., snow is white). In a paper writ-
ten several years later,16 Tarski again emphasised the relationship that exists
between statements in a formal language and the external world:
provide a new ‘way of explicating the concept of analyticity’ (Carnap 1952: 66).
The basic approach outlined in the paper involved sentences of the type ‘If Jack
is a bachelor, then he is not married’ being formally represented using meaning
postulates of the type
where, in this case, ‘B’ and ‘M’ indicate ‘is a bachelor’ and ‘is married’ respec-
tively. Since this implication is necessarily valid, Carnap suggested that the
following meaning postulate could be adopted
and the rest of the paper considered the ways in which adopting such postulates
can simplify the task of semantic analysis. As already implied above, this revival
of logical semantics in the 1930s and 1940s, which was largely engineered by
Ajdukiewicz, Tarski, and Carnap, can be viewed as a reaction against the more
extreme varieties of Formalism that attempted to exclude meaning entirely from
logical syntax.
To return to Bar-Hillel’s paper in the light of this summary, it is clear that
he was entirely familiar with the developments sketched above. For instance,
he refers to ‘the Warsaw-Lvov school’ explicitly, and mentions Ajdukiewicz
and Tarski by name. In addition, while considering these topics, he points to
similarities between the work of Bloomfield and Carnap concerning the role
of meaning. Since the relevant passage appears not to be well known, it merits
being quoted in full.
There are several issues here that merit comment. For instance, the parallel
between Bloomfield and Carnap is yet another instance of the association that
was felt to hold between logic and linguistics in the 1950s, and this helps to
explain why certain syntacticians borrowed techniques from formal logic (and
related disciplines) and used them to facilitate (or so they hoped) the analysis
of natural language. Bar-Hillel’s own advocation of logical syntax and recur-
sive definitions has already been considered, as has Chomsky’s adaptation of
constructional system theory. As Bar-Hillel acknowledges, one of the reasons
for this perceived closeness was the rejection of meaning that had characterised
Bloomfield’s and Carnap’s early work – an association that has already been
mentioned several times in earlier chapters (especially section 3.7). In the con-
text of this observation, it should be noted, though, that Bar-Hillel’s claim that
Bloomfield and Carnap worked ‘in complete independence’ is not entirely true.
As mentioned in section 3.7, Bloomfield had read LSL and in 1939 he wrote
an article for an Encyclopedia of which Carnap was one of the editors, though
admittedly it is difficult to gauge the full extent of their contact during this
period. Whatever the precise nature of their interaction, though, there is no
doubt that Bloomfield’s work seemed to make certain assumptions that were
similar to those made by Carnap in his pre-1935 publications, and, as noted
previously, to a later generation of linguists this apparent similarity appeared to
provide a basis for future developments. At the very least, it seemed to licence
a free borrowing of the techniques of formal logic for the task of linguistic
analysis.
While the picture painted by Bar-Hillel gives the impression of being clear
and comprehensive, it is (not surprisingly) rather too simplistic as an account
of the real situation. For instance, not all linguists shared his perception of this
close correspondence. As already indicated, Harris, for one, was convinced
that logicians and linguists did different things, and he was not forced into this
conclusion by ignorance of recent developments, for his familiarity with LSL
(at least) is well attested by the various references to Carnap’s book in MSL (as
discussed in section 3.7). Indeed, the difference between Harris’ and Bar-Hillel’s
views concerning logic and natural language could not be more striking, since,
according to the latter, linguists and logicians differ in ‘degree’ rather than in
‘kind’, while the former had stated that ‘linguists meet the problem [i.e., that of
132 Systems of syntax: 1951–1955
describing the structure of language] differently than do Carnap and his school’
(Harris 1952: 16.n17). The main difference between linguists and logicians
from Bar-Hillel’s perspective was that the latter were often more concerned
with the form of the systems they created than with the task of validating said
systems by means of empirical investigation. By contrast, linguists were forced
to compare the results of their theoretical investigations with utterances in actual
corpora. Therefore, issues such as ‘simplicity of handling’ (whatever that might
be exactly) and ‘ease of deduction and computation’ are primarily a concern for
logicians. However, as was shown in section 4.3, the notion of theory-internal
simplicity had already begun to infiltrate Chomsky’s conception of syntactic
theory largely due to his interest in Goodman’s work. Also, around this time,
Chomsky was starting to reconsider the usefulness of corpus-based discovery
procedures, and would eventually introduce a different kind of research method
that was not so deeply rooted in empirical validation. In both these respects, then,
Bar-Hillel’s description of linguistic investigation was soon to become outdated.
Nevertheless, the conception of linguistics and logic that he develops in his paper
is fundamental to Bar-Hillel’s view of their potential inter-connectivity.
As indicated above, Chomsky’s response to Bar-Hillel’s paper is of consider-
able interest for various reasons. His primary intention is to refute Bar-Hillel’s
suggestion that advances in the theory of logical syntax and semantics can
facilitate the analysis of meaning in natural language, a task of real significance
since, by the mid-1950s, Chomsky was fully persuaded that considerations of
meaning had no place in the study of syntax. For instance, to take just one
quotation from his other 1955 paper, ‘Semantic Considerations in Grammar’,
which also considers the role of meaning in linguistic theory, Chomsky had
plainly stated (in true Bloomfieldian fashion) that
from various critical accounts of the theory of meaning’ (Chomsky 1955a: 36).
More specifically, he explicitly names Quine’s LPV (especially chapters 2, 7,
and 8), as well as White’s paper ‘The Analytic and the Synthetic: An Untenable
Dualism’, which first appeared in 1950 (while Chomsky was still attending
White’s lectures?), but which was later published in the collection Semantics and
the Philosophy of Language edited by Leonard Linsky. The appearance of this
book in 1952 seems to have been partly responsible for provoking Bar-Hillel’s
and Chomsky’s exchange in the mid-1950s, since it is referred to in both papers
and raises many questions about the status of meaning in philosophical theories
of language. In particular, Quine’s and White’s attacks upon the traditional
Kantian distinction between analytic and synthetic truths, which were briefly
summarised in section 3.6, appear to have persuaded Chomsky, and his use of
their ideas in his own arguments implies his acceptance of their views. In this
context, the details of his argument will now be considered below.
Appropriately enough, Chomsky begins his paper by summarising Bar-
Hillel’s main points and immediately attacks the idea that Carnapian transfor-
mations can be profitably used by linguists in order to analyse natural language.
His central argument is that, since Carnap assumes such relations as ‘formal
consequence’ and ‘synonymy’ as primitives in the logical systems he creates,
his work offers no means of explaining or clarifying these notions, and therefore
the use of such relations cannot benefit linguistics except in a ‘trivial’ fashion. If
such relations were adopted in a linguistic theory, Chomsky claims, the valid-
ity of the resulting inferences could only be assessed by listing all possible
options as postulates of the given language system, an argument that Chomsky
had borrowed directly from White’s paper since, as mentioned in section 3.6,
White had argued that a ‘rule-book’ (White 1952: 277) of synonyms would be
required in order to deal with analytic truths in natural language, and (of course)
no such book exists. With considerable scorn, then, Chomsky declares that this
constitutes nothing more than ‘an ad-hoc approach to the problem of classi-
fication and characterisation of elements in particular languages’ (Chomsky
1955a: 38), and (in his opinion) the apparent failure of Carnap’s transforma-
tion rules to contribute anything of substance to the task of analysing natural
language phenomena such as synonymy entirely undermines Bar-Hillel’s trum-
peted belief in the efficaciousness of such rules in a linguistic context.
Having proceeded thus far in his demolition of Bar-Hillel’s argument, Chom-
sky pauses to consider the term ‘formal’, which ‘has played a rather crucial
role in this discussion’ (Chomsky 1955a: 39). In particular, Chomsky is keen
to reconsider the consequences of using such terminology while discussing
linguistic systems. For instance, Bar-Hillel had claimed in his paper that the
134 Systems of syntax: 1951–1955
active–passive relation that converts ‘plays’ into ‘is played by’ is a relation
of formal consequence. Chomsky does not accept this, since the meaning of
the term ‘formal’ has ‘misleading connotations’ in this context. Consequently,
he offers a (somewhat opaque) definition of ‘formal’ in which a relation is so
classified if ‘it holds between linguistic expressions’ (Chomsky 1955a: 39),
and while agreeing with Bar-Hillel’s statement concerning the active–passive
relation, he adds that the relation ‘longer by three words’ is also formal (in
the above sense), since it too holds for the expression pair ‘John did not come
home’ and ‘John came’. Since logical syntax of the Carnapian variety is unable
to determine which of these examples is an instance of formal consequence,
Chomsky is forced to conclude that relations of formal consequence are of no
use when analysing a natural language. Rather, the ‘systematic investigation
of linguistic expressions alone’ (Chomsky 1955a: 39) is required in order to
specify the sets of expressions for which such relations hold. Presumably, in
this context, the phrase ‘systematic investigation’ denotes some kind of Harris-
style distributional procedure that can be used to group expressions in a given
language, or possibly the type of constructional system-based approach devel-
oped in ‘SSA’. It is crucial to recognise, though, that such a procedure would
still be ‘formal’, since no semantic considerations would be permitted in such
a scheme. Chomsky concludes this section of his discussion with the unam-
biguous declaration that ‘logical syntax and semantics provide no grounds for
determining synonyms and consequence relations’ (Chomsky 1955a: 39).
The various arguments (summarised above) that Chomsky marshals in the
first part of his paper are partly designed to undermine Bar-Hillel’s claim that
linguistics could benefit from those aspects of Carnap’s work in logical systems
that deal with logical syntax. However, Chomsky is not convinced that logical
semantics has much to offer either, and, by rejecting this possibility, he reveals
the full extent of his dissatisfaction with contemporaneous accounts of linguistic
meaning. The causes of his dissatisfaction are established when he goes on
to assess the difference between logical implication and similar phenomena
in natural language. He considers several examples (including one borrowed
from Goodman) and easily demonstrates that natural language inferences are
unexpectedly complex, and therefore cannot be glibly analysed in terms of the
standard logical implication operator. Once again this provides him with an
ideal opportunity to stress the divide that separates natural language and the
type of artificial languages that Carnap creates in his ‘logical laboratory’.
(cited explicitly by Chomsky at the start of his paper) also served to destabilise
logical semantics. It is apparent, therefore, that Chomsky’s main intention in this
passage is twofold. First, he is keen to correct Bar-Hillel’s alleged misreading
of recent developments in logical semantics. However, second, he is also eager
to demonstrate that none of the recent advances cited by Bar-Hillel had enabled
meaning to be considered as a viable foundation for linguistic theory. Conse-
quently, as far as Chomsky is concerned, the post-Bloomfieldian emphasis on
the manipulation of meaning-less linguistic forms still provides the most secure
basis for syntactic theory, and, although this conclusion is not new, Chomsky’s
argument certainly demonstrates the extent to which his thinking on this issue
had been influenced by contemporaneous discussions in the field of analytic
philosophy, an influence that has not been adequately recognised in the past.
With his basic position now established, Chomsky seeks to render it more
secure by dismissing Carnap’s suggestion (endorsed by Bar-Hillel) that the
study of artificial languages can provide insights into the nature of natural lan-
guage. Consequently, Bar-Hillel’s assertion that logicians and linguists differ
only in ‘degree’ is shown to be mistaken: if one draws conclusions about nat-
ural language from logical systems constructed to explore the foundations of
mathematics ‘one might as well argue that a science-fiction writer or an artist is
doing roughly the same thing as a physicist’ (Chomsky 1955a: 42). The anal-
ogy is polemical, but Chomsky is able to conclude that ‘artificial languages
are neither special cases nor idealised versions of natural language’ (Chomsky
1955a: 42), and yet again the purpose here is to convey the impression that a
significant divide separates the disciplines of logic and linguistics, an attitude
that has already been identified in the work of Harris, suggesting that Chom-
sky was here dutifully following Harris’ lead. In general, Chomsky’s aim is to
acknowledge and emphasise the independence of linguistics as an intellectual
discipline. If the association between it and logic were accepted too rapidly
and too extremely, the former would simply be annexed to the latter, with (pre-
sumably) dire consequences. Once again, though, it is worth emphasising that
Chomsky himself had attempted to develop a system of linguistic analysis in
‘SSA’ that had concentrated on simple artificial languages, and the above pro-
nouncements must have been influenced to some extent by his own personal
experience of this type of approach.
In characteristic fashion, having arrived at the strident position outlined in
the above paragraph, Chomsky then immediately proceeds to reveal the full
complexity of his views. Although he remains adamant that ‘logical syntax
and semantics can bring the linguist no nearer to an adequate conception of
synonymy and transformations’ (Chomsky 1955a: 41), he does not rule out
Logic and linguistic theory 137
the possibility that logic can be fruitfully employed in a linguistic context, and
he specifically mentions Bar-Hillel’s work involving recursive definitions as a
recent positive example. Aware that such pronouncements could seem para-
doxical when juxtaposed with the previously articulated scepticism, Chomsky
seeks to clarify his position:
The correct way to use the insights and techniques of logic is in formulating
a general theory of linguistic structure. But this does not tell us what sort of
systems form the subject matter for linguistics, or how the linguist may find it
profitable to describe them. To apply logic in constructing a clear and rigorous
linguistic theory is different from expecting logic or any other formal system
to be a model for linguistic behavior. (Chomsky 1955a: 45)
This single passage reveals the full intricacy of Chomsky’s attitude towards the
use of techniques derived from formal symbolic logic in linguistic analyses,
and, despite some of the more extreme comments quoted above, it emerges
that he is neither simply for, nor simply against, the use of logic in linguistic
research. Rather, he is opposed to the unthinking assumption that logic will
necessarily provide insights into the structure of natural language. In the event,
some techniques derived from logic may prove to be useful when a linguistic
theory is being constructed, while others may offer no benefits. The task of
the linguist is sensibly to assess the validity of the various techniques in the
context of natural language analysis. As indicated above, Chomsky is seem-
ingly convinced that the methods of logic can certainly be usefully employed
when ‘a general theory of linguistic structure’ is developed, and, as with his
remarks concerning the use of artificial languages in linguistic research, his
comments here are not abstract musings since, during early 1955, Chomsky
was busy completing the first draft of LSLT, the main text in which he outlined
his own general theory of linguistic structure. The full title of this work alone
emphasises the point made at length above: The Logical Structure of Linguistic
Theory, with its conscious nod towards (the English translation of) Carnap’s
The Logical Structure of Language,17 implies that the structure of linguistic
theory must be logical. However, it does not follow from this that the task of
analysing natural language can be reduced in a trivial fashion to an exercise in
logical manipulation. In choosing his title, therefore, Chomsky seems simul-
taneously to be aligning himself with, and distancing himself from, Carnap’s
work. Consequently, it is appropriate that many of the insights into the nature
of the relationship between logic and linguistics, which can be gleaned from
his 1955 paper, can be considered in relation to various aspects of LSLT, some
of which will be explored in detail in chapter 5.
138 Systems of syntax: 1951–1955
for Montague, natural and logical languages were identical, or, as he expressed
it in the trenchant beginning of his 1970 paper ‘English as a Formal Language’,
‘I reject the contention that an important theoretical difference exists between
formal and natural language’ (Montague 1970: 188). Consequently, Montague
was able to develop a detailed formalism for a truth-theoretical analysis of
natural language which was explicitly intended to provide an alternative to
generative grammar, indicating that the old divisions of the 1950s were still
clearly motivating research twenty years later.18
5 Transforming generative
grammar: 1955–1957
140
Stochastic processes and autonomous grammar 141
the grammar of a language could be considered in isolation from all other aspects
of language (especially semantics). In section 4.5 the various arguments that
Chomsky developed in the mid-1950s to justify his rejection of logical semantics
were discussed at length, and one consequence of this was that, following the
tradition of Bloomfield and Harris, Chomsky became convinced that syntax
could be studied separately from semantics. In fact, he claimed specifically
that syntax provides a basis for semantics.1 However, Chomsky not only felt
compelled to rescue syntax from the hands of the logicians, but he was also
keen to preserve it from the clutches of statisticians. In particular, he responded
negatively to the concept of a stochastic grammar, which had been proposed
by Shannon and Weaver in the 1940s, and which had been enthusiastically
welcomed by certain post-Bloomfieldian syntacticians. Consequently, before
discussing Chomsky’s rejection of these ideas, it is necessary to summarise
them in some detail.
In the first decades of the twentieth century the statistical properties of nat-
ural language became an active area of research. For instance, Andrei Markov
(1856–1922) introduced stochastic processes, called finite state automata, and
used them to model letter frequencies in Russian poetry (see Sheynin 1988).
Also, George Zipf (1902–1950) published two books, The Psycho-Biology of
Language (1935) and Human Behavior and the Principle of Least Effort (1949),
in which he explored frequency counts for linguistic units such as phonemes,
syllables, and words, observing that they were invariably distributed in a char-
acteristic fashion, with a small group of frequently occurring units and a long
tail of infrequent units. This type of distribution, which can be represented as
a linear plot when the natural logarithm of the frequency is plotted against the
natural logarithm of the unit number, was later called a Zipfian distribution,
and research of this kind suggested that there were patterns hidden in natural
language that could best be revealed by detailed statistical analysis. In turn, this
implied that stochastic models of linguistic behaviour constituted a branch of
statistical theory that was well worth exploring. This general project was given
significant impetus by the publication of Claude Shannon (1916–2001) and
Warren Weaver’s (1894–1978) The Mathematical Theory of Communication in
1949.2 This text, which became one of the seminal works of information theory,
provided a clear introduction to a number of comparatively recent advances in
the theory of stochastic processes, thus making these techniques available to a
wide, non-specialist audience for the first time.
From the viewpoint of linguistic theory, the most significant mathematical
model discussed by Shannon was the finite state machine; that is, the stochastic
process introduced by Markov which, by the 1940s, had come to be known as
142 Transforming generative grammar: 1955–1957
Language must be designed (or developed) with a view to the totality of things
that man may wish to say; but not being able to accomplish everything, it too
should do as well as possible, as often as possible. This is to say, it too should
deal with its task statistically. (Shannon and Weaver 1949: 117)
Accepting that to do ‘as well as possible, as often as possible’ is all that syntactic
theory can hope to achieve, the suggestion here is that language must be studied
as some kind of stochastic information source, since only a model of this type
will permit the sort of approximations which (it is supposed) are required when
natural language is modelled mathematically. However, as will be discussed
below, this humble desire was not shared by the more idealistic members of
Stochastic processes and autonomous grammar 143
Once again, therefore, Hockett stresses his conviction that stochastic processes,
at least in theory, provide an adequate model for the grammar of a natural
language, since, as indicated above, he believed that ‘the entire grammatical
structure of English’ could be analysed using statistical techniques.
In the light of this support for stochastic techniques amongst certain lin-
guists, particularly Hockett, it is revealing to consider Chomsky’s assessment
of these techniques,3 and the most complete presentation of his ideas concern-
ing stochastic grammars can be found in his 1956 paper ‘Three Models for
the Description of Language’ (hereafter ‘TMDL’) which (it should be noted)
appeared in the I.R.E. Transactions for Information Theory. In fact, the paper
was delivered as part of a series that focused on various aspects of information
theory, and the event took the form of a conference at MIT in 1956. Such a gath-
ering was hardly a conventional forum for a discussion of the syntax of natural
language, and it is revealing that this particular conference should be consid-
ered by some to indicate the emergence of fully-fledged cognitive science.4 As
the title of his 1956 paper suggests, Chomsky’s main purpose was to explore
three different models (i.e., grammars) that provided different descriptions of
a given language. Specifically, the three types of model he considers are finite
state grammars, phrase structure grammars, and transformational grammars.
Various aspects of his presentation of the latter two types will be considered
in the other sections of this chapter, therefore, for now, the emphasis will be
exclusively upon his assessment of finite state grammars.
As expected, Chomsky’s presentation of finite state machines was derived
primarily from Shannon and Weaver’s 1949 book. Consequently, following
them, he states that a finite state grammar, G, for a given language, L consists
of
1. 1 ≤ i < j ≤ n
2. there are symbols bi , b j ∈ A with the property that S1 is not a sentence
of L, and S2 is a sentence of L, where S1 is formed from sentence S
by replacing the i th symbol of S (namely, ai ) by bi , and S2 is formed
from S1 by replacing the j th symbol of S1 (namely, a j ) by b j .
The above conditions rely upon the concept of replacement, which, though
not defined, is understood to correspond to substitution. In this way, the
type of technique Chomsky uses in order to investigate ‘the absolute lim-
its’ of finite state grammars, which exploit the distributional properties of
symbol pairs, are closely related to some of the methods employed by Har-
ris and other post-Bloomfieldians. The basic idea behind replacement is that
two symbols, ai and a j , are dependent if, when ai is replaced by bi (where
ai = bi ), a j must also be replaced, this time by the symbol b j (where a j = b j ).
To consider a specific example, given the sentence S = a1 a2 a3
a4 , where S ∈ L, if the following sentences can be obtained by means of
replacement,
S1 = b1 a2 a3 a4 (5.1)
S2 = b1 a2 a3 b4 (5.2)
S1 = a a (5.3)
S2 = a b b a
S3 = a a b b a a
The dependency set for L can be given as D = {(1, 2m), (2, 2m −
1), . . . , (m, m + 1)} for any m, and since D can contain more than any fixed
number, m, of terms, it violates condition 1 above.
In the light of Chomsky’s doubts concerning the usefulness of artificial lan-
guages when analysing natural language (discussed in section 4.5), it is of
interest that, as in ‘SSA’, he here makes full use of simple toy languages in
order to determine the limitations of finite state grammars. However, having
determined these limits, Chomsky is eager to explore these findings in relation
to natural language. Consequently, in order to prove that finite state grammars
are not sufficiently powerful to provide a valid grammatical model for a given
natural language, he simply needs to demonstrate that natural languages contain
mirror-image structures (for instance) such as those presented in (5.3) above.
He accomplishes this by considering structures of the type
if S1 , then S2 (5.4)
148 Transforming generative grammar: 1955–1957
and by observing that there is a dependency between ‘if’ and ‘then’ in (5.4), since
the clause indicated by S1 could itself contain structures of the form ‘if . . . then’,
thus producing an English sentence with the mirror-image property assessed
previously. Consequently, Chomsky concludes that English fails condition 1
above and that therefore ‘no finite-state Markov process that produces symbols
with transitions from state to state can serve as an English grammar’ (Chomsky
1956: 113).
Having demonstrated that finite state grammars are not sufficiently powerful
to generate all the possible sentences of English, Chomsky then dismisses the
idea that n th order approximations, of the type proposed by Shannon and Weaver,
can ever be used to generate the set of grammatical sentences in English. The
thrust of his argument is that grammaticality and frequency are not related
notions, and to indicate that this is so he considers the sentences
Chomsky claims that these sentences (which were later made famous in SS),
occur equally infrequently in English, yet sentence (5.5) is grammatical while
sentence (5.6) is not. Therefore, he is obliged to conclude that frequency reveals
nothing about grammaticality. The only possible conclusion, therefore, is that
n th order approximations must be rejected since ‘as n increases, an n th order
approximation to English will exclude (as more and more improbable) an ever-
increasing number of grammatical sentences, while it still contains vast numbers
of completely ungrammatical strings’ (Chomsky 1956: 116).
As ever, it is essential to emphasise the full complexity of Chomsky’s argu-
ment. Indeed, there are, in fact, general parallels between his rejection of
stochastic techniques in ‘TMDL’ and his rejection of certain tools from logic in
‘LSS’. In particular, just as Chomsky had advised linguists not to apply blindly
the methodology of logic while attempting to analyse the structure of natural
language, but rather to select and adapt specific techniques if they seem to be
useful, so his advice to statistically minded linguists was actually expressed as
caution rather than prohibition. Certainly, as indicated above, he was convinced
that the grammar of natural language simply cannot be comprehensively mod-
elled as a stochastic process (despite the initial claims of Hockett and others),
but this did not lead him to advocate banishing statistical methods from lin-
guistics entirely. Indeed, on the contrary, he freely confessed that the form and
structure of natural language could be usefully explored using statistics, but his
scepticism concerning syntactic modelling remained.
From discovery to evaluation 149
Given the grammar of a language, one can study the use of the language sta-
tistically in various ways; and the development of probabilistic models for
the use of language (as distinct from the syntactic structure of language) can
be quite rewarding . . . One might seek to develop a more elaborate relation
between statistical and syntactic structure than the simple order of approxi-
mation model we have rejected. I would certainly not care to argue that any
such relation is unthinkable, but I know of no suggestion to this effect that
does not have serious flaws. (Chomsky 1957b: 17.n4)
In the excised section above, Chomsky cites recent work by Herbert A. Simon
(1916–2001) (specifically, Simon 1955) and Benoit Mandelbrot (b. 1924)
(specifically, Mandelbrot 1954) as specific examples of revealing statistical
studies, and, as the above passage indicates, he is willing to accept that statisti-
cal analyses of linguistic structure can highlight significant patterns. However,
he emphasises that these studies are primarily descriptive in intention and do not
make elaborate claims about the actual mechanisms of language. For Chomsky,
these sorts of concerns must be left for the linguists to discuss, since linguists are
more inclined (than statisticians) to focus upon the actual properties of natural
language, and therefore they are better able to resist the temptation to become
obsessed with peripheral non-linguistic considerations.
a grammar from scratch given a specific corpus. This basic development is well
known and has been discussed extensively.5 However, Chomsky’s switch from
discovery to evaluation procedures has never really been adequately considered
in the light of his changing response to the kind of logical empiricism advocated
by Goodman (in particular). Consequently, these aspects of Chomsky’s work
will be discussed here.
As mentioned in section 4.4, Chomsky’s first published paper, ‘SSA’,
appeared to suggest that he was contentedly working within a post-
Bloomfieldian paradigm, since the point of that paper was to explore the pos-
sibility of utilising Goodman’s calculus of individuals as a mechanical dis-
covery procedure that could automate grammatical analysis. However, by the
mid-1950s Chomsky’s understanding of the general purpose and function of
linguistic theory had clearly begun to change with the result that, between the
appearance of ‘SSA’ in 1953 and the completion of the first draft of LSLT some
time in early 1955, the emphasis of his research had shifted away from tax-
onomic discovery procedures towards the type of theory that would later be
referred to as TGG. Chomsky’s own account of this change in direction is well
known but is still worth quoting in this context.
By 1953 I came to the same conclusion [i.e., as Halle]: if the discovery proce-
dures did not work, it was not because I had failed to formulate them correctly
but because the entire approach was wrong. In retrospect I cannot understand
why it took me so long to reach this conclusion – I remember exactly the
moment when I finally felt convinced. On board ship in the mid-Atlantic,
aided by a bout of seasickness, on a rickety tub that was listing noticeably – it
had been sunk by the Germans and was now making its first voyage after
having been salvaged. It suddenly seemed that there was a good reason –
the obvious reason – why several years of intense effort devoted to improv-
ing discovery procedures had come to nought, while the work I had been
doing during the same period on generative grammar and explanatory theory,
in almost complete isolation, seemed to be consistently yielding interesting
results. (Chomsky 1979a: 131)
Whatever the truth of this story (and whatever ‘the obvious reason’ might actu-
ally have been), the underlying causes of this general shift in perspective were
many and various, though, certainly, the fact that ‘several years of intense effort’
had been (seemingly) wasted on discovery procedures could have provided the
main motivation for the change. In addition, though, there is no doubt that the
prioritising of evaluation over discovery procedures was partly prompted by
Chomsky’s evolving appreciation of the work of Goodman and Quine. In partic-
ular, during the early 1950s the type of hard-line Carnapian logical empiricism
From discovery to evaluation 151
If these remarks can be trusted, then this passage is of particular interest since it
reveals a considerable amount about the underlying philosophical motivations
that prompted the development of TGG. Crucially, it identifies some of the
causes for the dramatic shift away from the sort of empirical discovery pro-
cedures that had characterised post-Bloomfieldian linguistics, and which had
provided the focus for Chomsky’s own early work. Specifically, it is perhaps
surprising that, in the above account of the development, Chomsky’s motivation
seems to have been primarily philosophical in origin rather than purely linguis-
tic. As was discussed in section 3.5, ever since Carnap had formulated construc-
tional system theory in the LCW, the use of such systems had been strongly
associated with Carnapian logical empiricism, which, as a theoretical approach
to the problem of knowledge acquisition, was (as the name suggests) funda-
mentally empirical, with objects in a given system (i.e., the ‘Gegenstanden’)
being ultimately derived, by means of definition, from sensory experience.
However, the above passage suggests (ironically) that Goodman’s attempts in
the early 1950s to justify the type of inductive procedures required by logi-
cal empiricism, combined with Quine’s probing scepticism about empiricism
152 Transforming generative grammar: 1955–1957
Understandably, then, more critical thinkers have suspected that there might
be something wrong with the problem we are trying to solve. Come to think
of it, what precisely would constitute the justification we seek? If the problem
is to explain how we know that certain predictions will turn out to be correct,
the sufficient answer is that we don’t know any such thing. If the problem is to
find some way of distinguishing antecedently between true and false predic-
tions, we are asking for prevision rather than for philosophical explanation.
(Goodman 1954: 62)
The clear implication of this overview is that inductive procedures are beset
with real difficulties and, to provide a contrast, Goodman then considers ‘non-
inductive inferences’ (i.e., deductive processes) and, in this case, the situation
is less complex.
Having outlined his basic concerns in this general fashion, Quine then launches
a blistering attack on logical empiricism, as practised by Carnap in LCW (and
Goodman in SA, and Chomsky in ‘SSA’), and his basic criticism is that the
whole approach was ‘left in a sketchy state’ (Quine 1953: 40) since Carnap
never adequately indicated how statements such as ‘Quality q is at point-instant
(x, y, z, t)’ could ever be translated into the parsimonious initial language that
consisted solely of logical axioms, extralogical primitives, and sense data. As a
result, the whole methodology was flawed. It is important to emphasise, though,
that Quine’s purpose in his paper was to propose a modification of empirical
philosophy, not entirely to reject it (‘As an empiricist I continue to think of the
conceptual scheme of science as a tool, ultimately, for predicting future expe-
rience in the light of past experience’ (Quine 1953: 44)), and, in this respect,
his position is similar to Goodman’s concerning induction: both men desta-
bilise aspects of empiricism and then proposed ways of rendering them more
secure. As indicated in this brief summary, the critique of induction and logical
empiricism that was developed in the work of Goodman and Quine would cer-
tainly have interested Chomsky in the early 1950s, given his experiments with
154 Transforming generative grammar: 1955–1957
The thrust of Chomsky’s argument, then, was that while the mechanical pro-
duction of a grammar by means of discovery procedures may be a desirable
accomplishment, it may simply be impossible to achieve, whereas a less ambi-
tious approach could actually provide ‘answers’. As Chomsky indicates, the
fundamental difference between the three positions can be simply stated as
the privileging of evaluation procedures (i.e., the selection of one grammar
from amongst many, given a particular corpus) over either discovery or deci-
sion procedures (i.e., either the automatic inferring of a grammar from a given
corpus, or the automatic assessment of the validity of a grammar), and the
advice is to focus upon the relativistic exploration of the validity of com-
peting grammars, rather than upon the notion of absolute correctness. As
noted in section 4.3, Goodman’s research into simplicity measures for con-
structional system theory partly motivated this reassessment of the linguist’s
task by considering the relative simplicity of the extralogical components in
various competing systems. In addition, it should also be obvious that the
task Chomsky came to advocate was the least empirical of the three tasks
defined; a change that can be viewed as a manifestation of his response
to Goodman’s and Quine’s critiques of induction and logical empiricism
respectively.
Whatever the precise nature of the influences that prompted his revisions,
there is no doubt that Chomsky actively set about redefining the goal of linguis-
tics (and especially syntactic theory) in LSLT, a work in which the rejection
of inductive discovery procedures is clearly stated. The avoidance of a purely
nominalistic constructional methodology is signalled early in the book when
Chomsky explains that, although such an approach is ‘more natural’ than the
analytic framework adopted in LSLT, it will not be further developed because
it necessitates the use of ‘somewhat more elaborate constructions’ (Chom-
sky 1975[1955]: 110); a remark that recalls Quine’s complaint concerning the
‘unnecessarily and intolerably restrictive’ (Quine 1953: 38) methods required
by the more extreme forms of logical empiricism. Nevertheless, despite this
unambiguous change of direction, it is equally apparent that LSLT is securely
rooted in constructional system theory. Indeed, it is a curious aspect of the
development of TGG that, although the basic philosophy of inductive logical
empiricism was spurned during the years 1953–1955, the technical devices
associated with constructional system theory continued to be utilised exten-
sively. Somehow, the rejection of the philosophy underpinning the theory did
not invalidate the tools it provided, and, as will be shown in section 5.4, Chom-
sky continued to use and develop the core formal procedures, adapting them to
the task of linguistic analysis.
156 Transforming generative grammar: 1955–1957
of morphemes and those that operate upon single morphemes, and both types
are presented as ways of incorporating information from outside the text into
the task of analysing a discourse. Harris’ intention was basically to define
equivalence classes in order ‘to discover patterned (i.e., similar or partly sim-
ilar) combinations of these classes in successive intervals of the text’ (Harris
1952: 18–19). Consequently, the notation Harris uses is different from that
which Chomsky had employed since, rather than employing a symbol that
denotes some kind of (quasi-)inference, Harris simply uses the equality oper-
ator to enable him to state equivalences. For instance, in Harris’ formalism,
the equivalence of active–passive pairs enables sentences such as ‘Casals plays
the cello’ to be considered equal to sentences such as ‘The cello is played by
Casals’, and, in formal notation, Harris defines this equivalence as follows:
N1 V N2 = N2 is V − en by N1 (5.10)
were not the same as Carnapian transformations, since, as mentioned above, the
latter are essentially rules of logical inference, while Harris’ statements cannot
obviously be classified as rules of this kind. Since it is known that Harris was
familiar with Carnap’s LSL, and since (presumably) he understood the structure
of the logical system presented in that text, it is either to be assumed that Harris
actually believed that his grammatical transformations defined rules of infer-
ence (which is unlikely), or else that he was consciously adapting a term taken
from logical syntax and redefining it within the context of linguistic analysis.
The latter alternative seems to be the most reasonable option, partly because (as
indicated in section 3.7) Harris was adamant that logicians and linguists studied
different things, and partly because (as stated above) Harris’ transformations are
not presented as if they were rules of inference. Also, revealingly, Harris never
cites Carnap as a direct source for his use of the term ‘transformation’. Indeed,
he generally implied that his own use of the concept was developed to facilitate
the analysis of utterances in natural language. For instance, writing in 1957,
he observed that ‘the study of transformations arose out of the attempt to con-
struct a method for analysing language samples longer than a sentence’ (Harris
1957: 283.n1), implying that a concern with discourse structure analysis pro-
vided the main inspiration. Obviously, this does not exclude the possibility that
he knowingly adapted terminology that he had first encountered in the logical
syntax literature, but his silence concerning this matter is curious nonetheless.
Bar-Hillel’s interpretation of Carnap’s transformation rules was reinforced
in Chomsky’s 1955 response (also discussed at length in section 4.5), provid-
ing further evidence to support the idea that, as far as linguists familiar with
recent developments in logic were concerned, Carnapian transformations did
indeed constitute rules of logical inference. For instance, while summarising
Bar-Hillel’s argument, Chomsky himself considers the nature of Carnapian for-
mation and transformation rules. He observes that, for Carnap, the notion of
formal consequence is given as a primitive in the system he constructs and,
while indicating that such an assumption would not contribute to the task of
analysing natural language, he comments that ‘the transformation rules of logic
are rules of valid inference; hence in this discussion the terms “inference” and
“transformation” will be used interchangeably’ (Chomsky 1955a: 37.n4). The
phrase ‘transformation rules of logic’ is revealing, since the implication is that
Chomsky was already thinking of linguistic transformations as being different
kinds of formal devices from those encountered in logical syntax, and it should
be remembered that, as mentioned above, the ‘transformation statements’ used
in MMH were certainly not strictly rules of valid inference. A slight confu-
sion occurs, though, when Chomsky goes on to consider Harris’ grammatical
Transforming transformations 163
The passage raises a number of issues that have already been discussed, and
it indicates how these issues were considered to be deeply interconnected at
the time. For instance, in the above passage we encounter the (by now famil-
iar) belief that techniques from logic can be usefully employed when natural
164 Transforming generative grammar: 1955–1957
We are thus led to develop a new level of syntactic analysis, the level T of
transformations, and to assign T-markers to strings of words as markers of their
‘transformational history’. That is, the T-markers of a string of words will tell
us how this string is derived from a certain kernel of sentences which have ρ1 -
derivations and P-markers. In terms of previous levels we can represent each
sentence as a string of phonemes, words, syntactic categories, and, in various
ways, as strings of phrases. Now we will be able to represent a sentence as
a sequence of operations by which this sentence is derived from the kernel
Transforming transformations 165
This passage emphasises the fact that, as already indicated above, the trans-
formations used in TGG operate upon existing strings of symbols (i.e., kernel
sentences) and modify their structure in precisely specified ways. It is the need
to restrict the applicability of the various transformations that leads Chomsky
to specify three basic conditions - namely, C1, C2, and C3 – which apply to the
set of grammatical transformations. These conditions can be defined as follows
(compare Chomsky 1975[1955]: 311):
I will try to show that a theory of transformations can provide a unified and quite
natural approach to all the problems mentioned in [earlier sections], and that
it can result in syntactic description which is considerably more economical
and revealing. (Chomsky 1975[1955]: 307)
As ever, the key word here is ‘economical’: in the event, Chomsky is keen
to demonstrate that the use of transformations greatly simplifies the proposed
grammar since it enables all types of recursion to be eliminated from the phrase
structure component. This important topic is discussed in greater detail in
section 5.6.
As the foregoing discussion indicates, the confusion concerning the nature
and purpose of grammatical transformations was widespread, and the relation-
ship between the transformations employed by Harris, Chomsky, Harwood, and
others, and those utilised by Carnap and his fellow logicians provided the main
focus for discussions of such topics throughout the 1950s. By the early 1960s
the situation seems to have become slightly clearer, mainly because Chomsky’s
detailed grammatical formalisms provided secure definitions of the terms as
used in the mature theory of TGG. For instance, writing in 1963, Bar-Hillel
indicated that, while he was personally convinced of the importance of TGG-
style transformation rules, he felt obliged to re-emphasise the fact that they were
not ‘transformation’ rules in the Carnapian sense, but rather ‘formation’ rules
(again in the Carnapian sense). Indeed, as far as Bar-Hillel was concerned, they
were a new kind of formation rule which permitted the analysis and descrip-
tion of ‘many more linguistic facts than I had originally thought’ (Bar-Hillel
1963: 542). However, as a result of the terminological confusion, Bar-Hillel felt
obliged to offer the following words of caution:
those formation rules that can handle these facts are of a type which has only
recently been analysed and understood . . . It is interesting that this novel
type of formation rules has been called by Harris and Chomsky (and is now
being called by everyone else) ‘rules of transformation’, and the reader should
beware of the confusion. (Bar-Hillel 1963: 543)
discussion, however, has been partly to revive something of the original chaos.
To summarise therefore, it seems likely that the transformations developed
by Harris and Chomsky were related to the transformation rules presented in
(the English translation of) Carnap’s LSL. However, since Harris’ grammatical
equivalences and Chomsky’s transformations are not obviously rules of valid
inference, they must be viewed as formation rules from a Carnapian perspective
(the perspective consistently adopted by Bar-Hillel). However, as Harwood’s
work shows, during the 1950s transformation rules came generally to refer
to rules that operated upon complete sentences that had been generated by
other rules, with the transformations causing modifications to the order of the
elements in the existing sentences. Obviously, it was this kind of transformation
that came to characterise Chomsky’s work in the years 1955–1957 (and beyond).
This passage indicates that, in mature TGG, the set of grammatical sentences,
L, is considered to be an infinite set of sentences, the elements of which can be
generated by the function, g, where g constitutes the grammar of the language
L. It is worth noting that (as indicated in section 3.3) even the terminology used
here reveals the connection between the theories: Post frequently used the verb
‘generate’ in order to describe how a recursively enumerable set is obtained from
its associated general recursive function, and Chomsky explicitly acknowledges
170 Transforming generative grammar: 1955–1957
Following a familiar technical use of the term ‘generate’ cf. Post (1944).
The locution has, however, been misleading, since it has erroneously been
interpreted as indicating that such sentence-generating grammars consider
language from the point of view of the speaker rather than the hearer. Actually
such grammars take a completely neutral point of view. Compare Chomsky
([1957b], p.48). We can consider a grammar of L to be a function mapping the
integers onto L, order of enumeration being immaterial (and easily specifiable,
in many ways) to this purely syntactic study . . . (Chomsky 1959a: 137.n1)
This passage indicates that Post’s work concerning recursively enumerable sets
was ‘familiar’ to Chomsky, and that the latter knowingly adapted the terminol-
ogy that the former had deployed. This fact should be borne in mind throughout
the following discussion since it is crucial to recognise that Chomsky was famil-
iar with Post’s own research and did not simply obtain his ideas concerning
recursively enumerable sets from later expositions of the theory.
The above discussion indicates that TGG, as it was presented in the 1950s,
required ‘recursive devices’ (Chomsky 1956: 116) in order to generate an infi-
nite number of sentences from a finite grammar, and, although Post provided
some kind of impetus for this assumption, Chomsky’s ideas concerning these
topics were also directly influenced by the work of Bar-Hillel. For instance, as
noted previously, Chomsky had remarked in ‘LSS’ that
1975[1955]: 194). However, Chomsky is quick to note that the finite nature
of a linear grammar requires the sequence of rules to be applied iteratively.
This passage indicates that the iterative application of the conversion statements
in the linear grammar was viewed (by Chomsky) as a recursive procedure, and
this type of recursion was explicitly introduced in order to enable an infinite
number of sentences to be produced by the grammar. In the light of the dis-
cussion of simplicity criteria in section 4.3, it is worth considering the recur-
sive application of conversion statements from a grammar-internal perspective.
Specifically, it is important to note that, for Chomsky, the issue of recursive rule
application was closely connected to the issue of determining the simplicity of
the resultant grammar, and the following passage is crucial:
As should be obvious, the formalism that Chomsky gives here is (to use Bar-
Hillel’s term) ‘recursive in disguise’ (Bar-Hillel 1953: 163) since he does not
provide the definitions of NP that are required in order to terminate the recursion,
yet these are provided in the main discussion of the topic that follows later in
the chapter. The basic point is clear, though, nonetheless: the phrase structure
component of the type of grammar developed in LSLT contains recursive rules.
In the above example, the rule indicates that NPs can be rewritten as larger
structures (containing relative clauses) which in turn contain NPs, and so on ad
infinitum, at least in theory if not in practice. The rest of chapter 7 constitutes an
exploration of the abstract theory underlying level P (i.e., the phrase structure
level of the grammar), and when Chomsky actually begins to develop this
component for English in chapter 8, he introduces a number of recursive rewrite
Recursive rules 173
rules. Sometimes these take the form of potentially infinite expansions. For
example, when discussing the structure of verb phrases, Chomsky proposes the
rules
Sentence → N P V P
V P → V PA V P1
V P1 → < V PB > V P2
V PA → V PA1 < V PA2 >
V PA1 → {C, ed} < M >
V PA2 → < have en >< be ing >
V PB → Z 1 < Z 2 < . . . < Z n >>>, where each Z i is one of the forms
Vc to, Vγ ing
(Chomsky 1975[1955]: 249–250)
Sentence
NP VP
N P V PA V P1
N P V PA V PB V P2
N P C V PB V P2
N P C want to V P2
J ohn C want to r ead the book
(Chomsky 1975[1955]: 250)
As these examples indicate, the recursive rules were initially introduced into
the phrase structure component of TGG; that is, they appeared in the part of
the grammar that was responsible for generating the kernel of basic sentences,
enabling infinite structures to be generated. However, as Chomsky begins to
elaborate his theory of transformations in chapters 9 and 10 of LSLT, he is
compelled to reduce the recursive elements in the phrase structure component.
More specifically, as the theory develops, the recursive rules are moved out of
the phrase structure component entirely, and are placed in the transformational
component of the grammar instead. In addition, Chomsky suggests that the first
type of recursion discussed above, which permits the conversion statements to
be applied recursively in sequence, should also be removed. Ultimately, these
developments led Chomsky to make the following claim towards the end of
chapter 10:
174 Transforming generative grammar: 1955–1957
In the course of this analysis we have found that much of the recursive part of
the grammar of phrase structure . . . has been cut away. It seems reasonable
to place the formal requirement that no recursions appear in the kernel gram-
mar. Specifically we rule out such statements as [V PB → Z 1 < Z 2 < . . . <
Z n >>>], and we drop the constructions . . . that permit running through
the grammar indefinitely many times. As far as I can determine, this formal
requirement on P does not exclude anything that we would like to retain in P;
nor does it impose any artificial or clumsy limitation on the actual statement
of the grammar corresponding to P, now that the transformational analysis
presents an alternative way of generating sentences . . . Now that the higher
level of transformational analysis has been established, it is no longer neces-
sary to require that generation by the grammar of phrase structure is infinite.
As the level T has been formulated, the process of transformational derivation
is recursive, since the product of a T-marker can itself appear in the P-basis
of a T-marker. (Chomsky 1975[1955]: 516–518)
The requirement that the phrase structure component should not contain recur-
sive elements ensures that the recursive parts of the grammar must all be imple-
mented in the transformational component; a development which, at the time,
no doubt emphasised the potential importance of Chomskyan transformations.
Nevertheless, as the above extract indicates, whether they are situated in the
phrase structure kernel or in the transformational component, TGG requires
recursive rules since such rules enable a grammar, g, to generate all the sen-
tences of the recursively enumerable language L, and, as suggested above, such
rules clearly have their roots in recursive function theory.
were incorporated into TGG at a comparatively early stage. However, the full
influence of Formalism upon the theory has yet to be illustrated.
As discussed at length in section 3.7, the need to formalise linguistic theory
became a pressing concern for particular groups of researchers in the late 1940s,
and this task had been partially accomplished by certain post-Bloomfieldians
(especially Bloch, Wells, Chao, Hockett, and Harris) by the mid-1950s. As
indicated previously, Harris was aware of contemporaneous developments in
logic and metamathematics and therefore it is no surprise that he encouraged
the young Chomsky to attend lectures on mathematics and philosophy. Once
again, therefore, although the following passage from Chomsky’s 1975
recollections was quoted in the introduction, it merits being reproduced here
since it provides a fascinating insight into the eclectic intellectual climate at
MIT in the early 1950s.13
research with Carnap’s research into logical syntax (i.e., in English translation,
Carnap’s The Logical Structure of Language), the main difference being that,
while Carnap had signalled his concern with the logical structure of language,
Chomsky was keen to stress his interest in the logical structure of linguistic
theories devised to analyse natural language. It is crucial to note the difference
in emphasis here: Chomsky’s title implies that, while it may not be possi-
ble to demonstrate that the structure of natural language can be reduced to
logic, it is certainly possible to ensure that the structure of a linguistic theory
is elaborated in a logical manner. Clearly, though, despite this difference in
emphasis, Chomsky seems to have been happy to associate himself (to some
extent) explicitly with the logico-philosophical Formalist tradition (as manifest
in Carnap at least), even if the association is intended primarily to illustrate
the distinctive character of his own approach; and, as the previous sections
have shown, such an association was a natural extension of the more formal
approach to the methodology of linguistics advocated by certain subsets of the
post-Bloomfieldians.
Given that developments in metamathematics were accessible to linguists in
the 1950s, and given that certain post-Bloomfieldians had sought to develop a
more systematic approach to linguistics by adopting a more (meta)mathematical
methodology, it is no surprise that, when TGG began to coalesce in the
mid-1950s, considerable emphasis was placed upon its status as a formal
theory. Indeed, Chomsky addresses this very issue in chapter 2 of LSLT,
and the Formalist nature of the project he is in the process of outlining is
clear.
This passage (or at least sections of it) could have been taken from any of the
textbooks discussed in section 2.8, and clearly the basic approach to syntactic
theory proposed here is entirely in keeping with Bloomfield’s recommended
axiomatic-deductive approach to theoretical linguistics. Indeed, the foregoing
sections of this book should already have illuminated the way in which the
above passage is richly embedded in a specific cultural tradition – namely, the
drive towards greater ‘rigour’ that was revived by Cauchy in the nineteenth
century in order to secure the foundations of the calculus, but which spread
to many other scientific disciplines during the twentieth century, largely due
to the influence of Logicism and Formalism. Before continuing, though, it is
necessary to emphasise Chomsky’s attitude towards the task of formalising a
linguistic theory.
As the above passage indicates, for Chomsky, formalisation for the sake of
formalisation is not the point: any theory can be converted into a ‘formalized’
theory, but this does not guarantee that the theory is ‘enlightening’. Conse-
quently, the motivation for propounding a formal linguistic theory is not to
spare the linguist the arduous task of analysing complicated and confusing data;
rather, the process of formalisation is desirable primarily because it enables spe-
cific conjectures concerning particular data to be stated more precisely. Indeed,
for Chomsky, formalisation was never a soulless activity that could be accom-
plished only when a theory was already complete; rather, it was a fundamental
aspect of the process of theory development. Indeed, as he observes elsewhere in
LSLT, ‘formalization can play a very productive role in the process of discovery
itself’ (Chomsky 1975[1955]: 59).
In the passage concerning formalisation quoted at length above, Chomsky
refers only in very general terms to the use of ‘axioms and methods of proof’.
However, the axiomatic nature of TGG is clear from the various presentations of
the basic theory that appeared in the mid-1950s. Once again, though, it is neces-
sary to make a basic distinction between the way in which Chomsky presented
his expositions of the basic theory, and the way in which language was analysed
within that theory. For example, the main chapters of LSLT conspicuously reveal
the axiomatic-deductive nature of Chomsky’s exposition. To take one example
virtually at random, when Chomsky introduces the ‘relation of representation’,
ρ, which he uses to specify relations such as ‘ρ(N P, the man)’, indicating
that ‘the man’ is an N P, he provides specific axioms and definitions, and
178 Transforming generative grammar: 1955–1957
then proceeds to deduce theorems from these. For instance, the axioms and
definitions take forms such as14
and the theorems are constructed using these axioms and definitions in the
standard axiomatic-deductive fashion. As is obvious even from these two
examples, unlike Bloomfield and Bloch (discussed in section 3.2), Chomsky
was not content solely with natural language descriptions of his various
technical entities and procedures; consequently, he introduces a rich symbolic
metalanguage that enables the various axioms and definitions to be presented
as precisely as possible.
While the presentation of TGG as a linguistic theory evinces a predilection
for the axiomatic-deductive method, it is important to emphasise the fact that
the basic sentence construction mechanism, as viewed from the perspective of
TGG, can also be classified as an axiomatic-deductive procedure. To consider
just one presentation, the rudiments of the theory were discussed explicitly
in ‘TMDL’, and, in this presentation, Chomsky seems to promote a distinctly
axiomatic-deductive interpretation of his theory since he describes the core
components of the phrase structure component as follows:
(i) Sentence → NP + VP
(ii) NP → T + N
(iii) VP → Verb + NP
(iv) T → the
(v) N → man, ball, etc.
(vi) Verb → hit, run, etc.
Formal syntax 179
Linguistic theory has two major subdivisions, syntax and semantics. Syntax
is the study of linguistic form . . . Semantics, on the other hand, is concerned
with meaning and reference of linguistic expressions . . . Syntax and semantics
180 Transforming generative grammar: 1955–1957
are distinct fields of investigation. How much each draws from the other is not
known, or at least has never been clearly stated. The subject of investigation
in the following pages will be syntactic structure, and we shall study it as
an independent aspect of linguistic theory . . . In part, our desire to place
no reliance on meaning in systematic development is motivated by a feeling
that the theory of meaning fails to meet certain minimum requirements of
objectivity and operational verifiability. (Chomsky 1975[1955]: 57)
This argument essentially restates that already presented in ‘TMDL’, and, once
again, the influence of the logico-philosophical Formalist tradition is clear. For
instance, Chomsky’s celebrated examples and general argument in the above
passage seem consciously to echo those introduced by Carnap in his LSL. As
mentioned in section 3.7, Carnap constructed the sentence ‘Pirots karulize elati-
cally’ in order to demonstrate that well-formedness could be analysed in purely
syntactic terms without reference to semantic considerations, and, obviously
Formal syntax 181
Chomsky is employing a similar strategy in order to make the same basic point.
Indeed, it is worth noting that Carnap’s example is actually more extreme than
Chomsky’s since the use of neologisms complicates the situation even more
from a semantic perspective, but there is no doubt that the main thrust of both
arguments is essentially the same.
It is clear that the type of argument summarised above is understood to apply
to syntactic research at a very general level; that is, the main recommendation
is that any sensible and self-respecting syntactic theory will distance itself
entirely (if possible) from considerations of meaning. However, while it is
sufficiently important to consider this kind of general recommendation, it is
equally revealing to explore the way in which Chomsky handles the problem of
meaning in his detailed developments of TGG. To give just one example of this,
in LSLT, when he is in the process of motivating the use of transformations in
addition to the conversions in the phrase structure component of the grammar,
Chomsky states:
The emphasis of this passage should already be familiar given the previous
discussion: Chomsky wishes to ensure that the transformations do not need
to take semantic information into account, and that they should consider only
the ‘number and order’ of substrings, since only a system of this kind will
enable the transformations to reflect ‘general structural relations’. The impli-
cation of this (and similar passages scattered throughout LSLT and SS) is clear:
Chomsky assumed that syntactic form could be exhaustively analysed inde-
pendently of meaning and, in assuming this, the type of theory he proposed
in the mid-1950s reveals something of its Formalist inspiration. At the very
least, it demonstrates how far syntactic theory had come since Hjelmslev dis-
tinguished between meaning-less metamathematics and meaning-full natural
language analysis in the 1940s. Indeed, Chomsky’s position is more extreme
even than that of second-generation Formalists, such as Carnap in the 1930s,
since, as indicated above, Carnap maintained that, due to their complexity, nat-
ural languages could not be analysed adequately using techniques that were
182 Transforming generative grammar: 1955–1957
183
184 Conclusion
If Bloomfield and Bar-Hillel have been dusted off a little in this book, then
perhaps a third rehabilitation has also been achieved – or perhaps even a third
and a fourth, depending on whether Goodman and Quine are sufficiently strong
collocators to count as a single entity or not. However that may be, one of the
aims of this research has been to explore the nature of the influence that the work
of Goodman and Quine exerted upon the development of TGG, particularly
since this topic has been so often neglected. As mentioned earlier, this neglect
is rather curious, since Chomsky himself has invited a consideration of these
issues. In the preface to SS, for instance, after dutifully thanking Harris for
his assistance, he adds that ‘in less obvious ways, perhaps, the course of this
research has been influenced strongly by the work of Nelson Goodman and
W. V. Quine’ (Chomsky 1957b: 6). Therefore, to this end, certain sections of
this book have simply explored this influence in greater detail than has been
accomplished in the past, and, as a result, various aspects of TGG have been
(re)assessed. For example, the use of simplicity criteria in Chomsky’s earliest
publications can be viewed as a direct adaptation of Goodman’s research into the
basal simplicity of constructional systems. Further, it has been demonstrated that
‘SSA’, Chomsky’s first published paper, is clearly an attempt to fuse Harris-style
discovery procedures with the particular constructional system that Goodman
had developed in SA. These early (i.e., pre-c. 1954) projects suggest that, at
this stage, Chomsky was still viewing linguistics as an empirical science that
required the elaboration of automatic discovery procedures. However, at some
point during the period 1953–1954 he appears to have rejected some of the
philosophical implications associated with the type of constructive nominalism
he had learnt from Goodman (and, to a lesser extent, Quine), and one of the
many consequences of this rejection was the rejection (in turn) of the use of
discovery procedures in linguistic theory in favour of evaluation procedures
designed to choose between competing grammars (or were the rejections in
the inverse order? or were they effectively simultaneous?). As indicated earlier,
this profound shift in Chomsky’s understanding of the purpose and function
of linguistic methodology has never really been assessed before in the light
of his changing attitudes towards constructive nominalism. Since Chomsky is
universally categorised these days as a self-professed ‘mentalist’, it is sometimes
difficult to accept that the pre-c. 1954 Chomsky seems to have been a logical
empiricist of the Carnap–Goodman variety, but the evidence contained in his
first publications and manuscripts certainly seems to point towards this con-
clusion. However, it should be emphasised that, even if Chomsky’s rejection
of logical empiricism was involved in his rejection of discovery procedures,
the evaluation procedures (i.e., simplicity criteria) that he recommended as
186 Conclusion
a preferable alternative also have their roots in the work of Goodman and
Quine. Indeed, it is a strange fact that, while Chomsky apparently came to
view constructive nominalism as having a deleterious effect upon linguistics,
the same theory nevertheless provided technical procedures that were involved
in his redefinition of the methodology of syntactic analysis. This conclusion
is admittedly rather complex, but it is to be hoped that the complexity offered
here is closer to the truth than the simplifications (or, worse, eerie silences) that
have characterised historiographical studies of the genesis of TGG in the past.
In addition to the three broad topics summarised above, a number of smaller
insights have been gained. For instance, with reference to non-TGG topics,
some forgotten linguists have been resuscitated, most notably Harwood, whose
early attempt to provide a formal axiomatic approach to syntactic theory seems,
unfortunately, to have fallen out of the collective consciousness. With reference
to TGG, though, the influence of White’s critique upon Chomsky’s rejection of
methods from logical semantics has been explored in a preliminary fashion, and
this association has not been fully recognised previously. Indeed, the treatment
of this topic (i.e., the influence of Quine–White style analytic philosophy from
the early 1950s upon TGG) is only sketchily accomplished in this book, and
the White–Chomsky relationship in particular still awaits comprehensive con-
sideration. In addition, there is the influence of Carnap’s research into logical
syntax upon the development of TGG, an association that has been generally
acknowledged in the past, but never discussed in depth. Specifically, in this
book, the difference between Carnapian and Chomskyan transformations has
been investigated, and it is hoped that some of the terminological complexities
have been elucidated. In general, though, perhaps the main achievement of
this book has been to associate TGG with both Formalism and Logicism, two
intellectual movements that profoundly influenced scientific methodology in
the early twentieth century. Indeed, this general issue seems to have been the
single destination towards which the various paths of enquiry have led. With
its focus on syntax as opposed to semantics, with its use of a logic-based
notational system, with its identification of the analogy between a proof and
the generation of a grammatical sentence, and with its use of such procedures
as recursive definitions and axiomatic deduction, TGG unambiguously reveals
its associations with the formal sciences.
Having summarised the main topics that this book has addressed, it is nec-
essary briefly to emphasise the main issues that have not been presented here.
In particular, it should be recalled that, as stated in the introduction, this study
of the development of TGG effectively ceases when the year 1957 has been
reached (this being the year in which SS was published), and one reason for
Conclusion 187
choosing this date as the terminus of the current study is to illustrate the void at
the core of the various discussions of the history of TGG that take 1957 as their
starting point. However, as a result of adopting 1957 as an upper-bound, there
has been no attempt in the present study to discuss Chomsky’s research into
formal language theory, which constituted one of his main academic preoccupa-
tions between the years 1958–1965. In particular, the work he published jointly
with both Miller and Marco Schützenberger (1920–1996) has not been assessed,
although such a re-evaluation is obviously necessitated by the various topics
addressed in the foregoing chapters. Similarly, with 1957 as the chosen cut-off
point, there is no space here for discussing the emergence of the innateness
hypothesis in exhaustive detail, the reason being that, although this hypothesis
was eventually to become one of the most characteristic features of generative
grammar, the initial expositions of this hypothesis were not presented until the
late 1950s and early 1960s. Given the subsequent dominance of this hypothesis
in the generative grammar tradition, this work clearly merits reconsideration
(in particular, Chomsky’s interactions with Eric Lenneberg (1921–1975) await
discussion), but such a task is outside the scope of the present study. In addition
to this nexus of interrelated topics, there are fascinating sociopolitical aspects of
the development of TGG that demand further exploration. For example, in the
late 1950s and early 1960s research into generative grammar was often funded
by the US defence department. However, as the 1960s advanced, such funding
sources ceased to be available for such work, although other kinds of research,
such as automatic speech recognition and machine translation, saw no diminu-
tion in their financial support during these years. The reasons for this shift in
perception from the institutions that initially financed TGG research have never
been discussed, barely even mentioned; and, alas, they remain neglected in this
book too, although the omission has (at least) been acknowledged. The ques-
tions are enticing, though. For instance, surely it can only be a coincidence
that the defence department funding of generative grammar research should
diminish in the late 1960s; that is, around the time that Chomsky started to
promulgate political views that were severely critical of US foreign policy.
Having provided this overview of the genesis of TGG in the light of various
developments in the formal sciences, it would be possible to terminate the dis-
cussion here, safe in the realm of conventional, retrospective historiography.
However, since it is always necessary to resist the artificial sense of complete-
ness that is inevitably associated with the conclusion of an investigation of
this kind, the discussion will be extended a little longer. The motivation for
continuing can be simply explained, and (perhaps predictably) it is prompted
by contemporary preoccupations concerning the nature of generative grammar.
188 Conclusion
Possibly such considerations seem out of place in a study of this kind, but it
should be remembered that, although TGG in its initial guise has long passed
into the historical twilight, more modern variants of the theory are still very
much in the full glare of daylight. Indeed, despite the many changes that have
occurred since the late 1950s (i.e., the influential reconstruction(s) of the 1960s
and 1970s that came to be referred to as the Extended Standard Theory (EST),
the advent of the Principles and Parameters (P&P) model in the 1980s, and
the subsequent inauguration of the controversial Minimalist Program (MP)),
the essential elements of the syntactic theory broadly known as ‘generative
grammar’ have remained remarkably constant. Consequently, certain aspects
of the theory that were originally bold assumptions (some of which were
ultimately derived from the formal sciences) have crystallised and have become
incontrovertible components. As ever, though, consequences eventually follow
from assumptions, and it is worth pausing to assess some of the implications
that are inherent in the framework adopted by generative grammar.
One aspect of TGG that has been emphasised throughout the foregoing chap-
ters is the conspicuously axiomatic-deductive (quasi-)proof-theoretical char-
acter of the theory. As demonstrated in sections 3.2 and 5.7, the axiomatic-
deductive method fascinated certain linguists during the first half of the twen-
tieth century, and it clearly provided the conceptual foundation for the type of
computational syntactic theory that Chomsky outlined in LSLT. Indeed, to take
this discussion back to where it (almost) began, it should be recalled that Lees
had noted, in his 1957 review of SS, that Chomsky’s theory utilised ‘an overt
axiom system’ (Lees 1957: 378), suggesting that this was one of the characteris-
tic features of the theory that made it ‘rigorous’ (Lees 1957: 378); and certainly
the axiomatic-deductive framework (with all its attendant logico-philosophico-
mathematical baggage) has remained a fundamental component of generative
grammar ever since. In his 1995 presentation of the MP, for example, Chomsky
sketched the basic structure of C H L , the computational component of Universal
Grammar (UG), and the proof-theoretical nature of this system is clear. For
instance, recursion is still explicitly utilised (i.e., ‘the operations of C H L recur-
sively construct syntactic objects’ (Chomsky 1995: 226)), and derivations are
‘generated’ in a derivational (as opposed to a representational) fashion, start-
ing with primitive elements (i.e., ultimately lexical and formal features) and
resulting in phonological form and logical form pairings (i.e., (π, λ)) (Chom-
sky 1995: 219ff.). Although it is not possible exhaustively to consider all aspects
of the various proof-theoretical features of contemporary generative grammar
here, it is revealing to probe the implications of certain assumptions that have
been retained; and, to this end, the formulation of recursion within contemporary
Conclusion 189
Chomsky explicitly notes that clause 2 of definition 6.1 provides the ‘recursive
step’ (Chomsky 1995: 243):
Suppose a derivation has reached state = {α, β, δi , ..., δn }. Then application
of an operation that forms K as in [definition 6.1, clause 2] converts to
= {K , δi , ..., δn } including K but not α, β. (Chomsky 1995: 243)
This summary is rather brief, and it is useful to work through a specific example
in order to appreciate the framework that is being proposed. For instance, if it is
assumed that the only operation that creates K is Merge, then derivation creation
can be viewed primarily as a process involving the repeated application of this
operation, which terminates when the initial numeration is mapped to a single
syntactic object. Schematically, if α1 , α2 , α3 , and α4 are the lexical items in a
given numeration, then for a derivation that begins in state = {α1 , α2 , α3 , α4 },
one possible series of subsequent steps can be explicitly represented as
follows:
Derivation: Example 1
Given = {α1 , α2 , α3 , α4 }:
Step 1: K 1 = Merge(α1 , α2 ) and = {K 1 , α3 , α4 }
Step 2: K 2 = Merge(K 1 , α3 ) and = {K 2 , α4 }
Step 3: K 3 = Merge(K 2 , α4 ) and = {K 3 }
The Recursion Convention has brought ‘recursive’ to have at least four differ-
ent meanings . . . This leads to some ambiguity . . . Worse still, the Convention
leads to imprecise thinking about the basic concepts of the subject; the term
‘recursion’ is often used when the term ‘computability’ is meant. (By the term
‘recursive function’ does the writer mean ‘inductively defined function’ or
‘computable function’?) Furthermore, ambiguous and little recognized terms
and imprecise thinking lead to poor communication both within the subject
and to outsiders, which leads to isolation and a lack of progress within the
subject, since progress in science depends upon the collaboration of many
minds. (Soare 1996: 29)
Chomsky, and Fitch are explicitly associating ‘recursion’ with the process of
creating an infinite arrangement of discrete expressions from a finite set of
elements (Hauser, Chomsky, and Fitch 2002: 1571, quoted above). This is
reasonable, though it is essential to note that this is a far more general procedure
than that associated with other interpretations of ‘recursion’. For instance, if this
elusive term is understood to mean ‘definition by induction’ (the interpretation
that Bar-Hillel recognised in 1953 when he proposed that recursive definitions
could be used in formal grammars), then in addition to the property of generating
infinite structures from finite means, ‘recursion’ would also involve some kind of
explicit self-reference. Since self-reference is not stressed in their paper, though,
Hauser, Chomsky, and Fitch may be using the term ‘recursion’ to indicate
something like ‘effectively calculable’ or ‘specified by a finite algorithm’, and
if this were the case, then it would suggest that Chomsky’s current use of the
term is more closely related to the interpretation of ‘recursion’ that associates it
with computability theory rather than with Gödelian recursive function theory.
However, as mentioned previously, since the presentation offered in the Hauser,
Chomsky, and Fitch paper is both brief and informal, it is simply not possible to
determine with certainty which interpretation of ‘recursion’ is actually intended.
Clearly, this is not the place in which to attempt to resolve some of these
problems, and the preceding discussion is presented merely to indicate that it is
impossible insightfully to assess the linguistic/biological role of the recursive
components within contemporary generative grammar without first reflecting
upon the tortive development of recursive function theory and its complex
associations with λ-calculus and computability theory; and explorations such
as that in section 3.3, which attempt to identify the way in which recursive
definitions were incorporated into syntactic theories in the 1950s, simply seek
to provide some kind of clarification.
If recursion is one topic that requires careful and dextrous consideration,
then the notion of theory-internal simplicity is certainly another. As discussed
(possibly ad nauseam) in sections 3.5 and 4.3, the characteristic concern with
simplicity criteria in early TGG was directly motivated by Chomsky’s inter-
est in Goodman’s research into the basal simplicity of constructional systems,
and related concerns have haunted generative grammar ever since its inception.
From the mid-1960s onwards, these concerns were often discussed in terms
of explanatory, as opposed to descriptive, adequacy, and the opposing tensions
caused by these guiding principles are well known: while descriptive adequacy
requires the full complexity of those aspects of natural language that fall within
the specified domain of enquiry to be captured by a formal linguistic theory,
explanatory adequacy essentially demands that the theory developed should not
194 Conclusion
become unnecessarily complex. Due to the fact that the introduction of explana-
tory adequacy as a theoretical prerequisite roughly coincided with the explicit
adoption of the innateness hypothesis, the notion has conventionally been pre-
sented as a psychological/biological requirement in the generative grammar
literature. In general, then, during the years c. 1965–1990 the main focus in
mainstream generative grammar research was upon the task of improving the
basic theory so that it could account more accurately for a larger number of lin-
guistic phenomena, and the task of further developing the ‘sketchy’ ideas con-
cerning simplicity criteria outlined in LSLT (Chomsky 1975[1955]: 116) was
largely abandoned. Nevertheless, during this 25-year period of theory develop-
ment, a basic sense of the importance of theory-internal simplicity was retained,
and there were certainly times when the eradication of certain unnecessary the-
oretical complexities was openly acknowledged as a profound advance. For
example, the development of X-bar theory in the early 1970s, which consid-
erably simplified the orthodox theory of generative grammar by providing a
unified analytical framework for phrasal projections, was widely recognised
as a significant improvement. However, despite such cases, it is reasonable to
claim that, in pre-c. 1990 generative grammar, considerations of ‘economy’ or
‘theory-internal simplicity’ never received the type of principled and sustained
scrutiny that they began to receive when the MP emerged as an identifiable
research programme in the early 1990s.
As is well known, the MP outlines an ambitious agenda, the main purpose
of which is to determine the extent to which the human language faculty can
be viewed as a ‘perfect’ solution to the problems posed by external constraints
(i.e., the interfaces that connect the syntactic component with the phonological
and semantic components). Put simply, one of the main goals of the MP is to
reduce to a minimum the essential machinery required to represent the faculty
of language, or, as Chomsky himself has expressed it recently,
The minimalist program is the attempt to explore these questions [i.e., the
questions ‘what are the properties of language’ and ‘why are they that way’]. Its
task is to examine in detail every device (principle, idea, etc.) that is employed
in characterizing languages to determine to what extent it can be eliminated in
favor of a principled account in terms of general conditions of computational
efficiency and the interface condition that the organ must satisfy for it to
function at all. Put differently, the goal is to determine just what aspects of
the structure and use of language are specific to the language faculty, hence
lacking principled explanation at this level. (Chomsky 2004: 106)
In the terms of the MP, then, the concern with issues of theory-internal econ-
omy is formulated as a set of economy considerations that prevent non-optimal
derivations converging, and (as indicated in the above quotation) Chomsky him-
self acknowledges that there is a noticeable similarity between this approach
and the TGG focus upon simplicity criteria that selected between competing
grammars. The implication is that the MP can be viewed (in part) as an attempt
seriously to respond to certain questions concerning the theory-internal ‘sim-
plicity’ of formal grammars that were asked in early work on TGG, but which
have remained largely unanswered during the intervening decades. The fact that
such concerns have ‘resurfaced’ is perhaps surprising, but it certainly indicates
that research concerns that were responsible for prompting the emergence of
TGG in the 1950s are closely related to the research concerns that prompted
the advent of the MP, and such associations surely merit detailed considera-
tion. Obviously, the task of determining the nature of theory-internal economy
is non-trivial, and this is indicated by the various ways in which the topic is
196 Conclusion
has good reason for being insecure because he cannot fail to have noticed that
he has few substantial results in the sense that these are understood in more
Conclusion 199
This pronouncement is unambiguous, and uttered only three years after Postal’s
contemptuous assessment. So which analysis is correct? Is generative gram-
mar a science or not? The divergence between Postal and Piattelli-Palmarini’s
views is rather alarming to say the least. Surely it should not be possible for
the opinions of informed, practising linguists to differ so greatly on this mat-
ter? Surely it should be possible for the linguistics community to agree as to
whether generative grammar has indeed produced any genuine results or not?
Surely it should at least be possible to determine whether, after over half a
century of formalisms, debates, analyses, assertions, counter-assertions, mis-
understandings, revelations, and revolutions, the researchers involved with the
generative grammar enterprise (whether TGG or EST or P&P or the MP) have
actually been responsible for some kind of scientific research or not? And if
such possibilities are not, in fact, possible, then surely it is the case that some-
thing is rather deeply wrong with syntactic theory, or at least with the particular
branch of it known as generative grammar? Unfortunately, there is not space
here fully to explore all the implications raised by these queries. However, if
200 Conclusion
Introduction
1 For short but insightful discussions of the standard classifications (and some of the
attendant difficulties) see Suppe 2000 and Schaffer 1997.
2 For a notorious negative review of Chomsky’s book see Aarsleff 1970. For a more
dispassionate assessment, and for evidence of Chomsky’s annoyance at the reception
of Cartesian Linguistics, see the discussion in Koerner 1999, especially pp. 10–12,
p. 178, and pp. 210–214.
3 Some of the mathematical aspects of Hockett’s own work are discussed in
section 5.2.
4 Chomsky’s rejection of nominalism is discussed at length in section 5.3. For Lyons’
discussion of Chomsky’s critique of behaviourism, see 1970: 83ff., and for Chom-
sky’s own words, see Chomsky 1959b.
5 For the details see Matthews 1993: 131–134.
201
202 Notes to pages 27–38
8 See section 2.5 below for a brief summary of the development of logic in the nine-
teenth century.
9 The most authoritative overview of Cauchy’s contribution to the nineteenth-
century reformulation of analysis is Grabiner 1981. The development of Cauchy’s
ideas in relation to logic and set theory is considered in Grattan-Guinness 2000:
64–68.
10 Good surveys of various aspects of Weierstrass’s work can be found in Dugac
1973 and Manning 1975. As an aside, it should be mentioned that the consensus
amongst the mathematical community concerning the definitions of differentiation
and integration was undermined when Abraham Robinson introduced Non-Standard
Analysis in the 1960s. This theory provided a model-theoretic basis for Leibnizian
infinitesimals, therefore reviving the very conceptual foundation for the calculus that
Weierstrass had effectively eliminated. For further information about this fascinating
development, see Robinson 1996 and Dauben 1995.
11 A good explanation of ‘Dedekind cuts’ can be found in Hrbacek and Jech 1984: 100–
102.
12 An excellent introduction to Cantor’s research into number theory can be found in
Dauben 1979, especially chapter 2.
13 For more detailed accounts of Cantorian set theory, see Dauben 1979 (chapters 10
and 11), Jech 1991, and Grattan-Guinness 2000 (chapter 3). For a brief but insightful
summary, see Hallett 1984: 1–11.
14 For more information about set theory in the twentieth century, see Johnson
1972.
15 For general overviews of the development of logic, see Kneale and Kneale 1962 and
Grattan-Guinness 2000, chapter 4.
16 For more information about Frege’s logical systems see Baker 1984 and Dummett
1991. As the twentieth century progressed Frege was gradually appropriated by
analytical philosophers. For a recent attempt to reclaim him for logicians and math-
ematicians, see Grattan-Guinness 2000: 177–199.
17 By comparison with the work of other logicians concerned with foundational issues,
Peano’s work has been rather neglected. For a reasonably detailed overview, see
Grattan-Guinness 2000, chapter 5.
18 For more information concerning Russell’s intellectual development, see Clark 1975
(chapter 5), and Hylton 1990 (chapter 3).
19 In the years following the publication of Whitehead and Russell’s ideas concerning
the theory of logical types, a number of criticisms of the theory emerged. For a
general overview of the theory itself and the controversy surrounding it, see Copi
1971.
20 For a more detailed discussion, see Grattan-Guinness 2000, especially chapter 7.
21 For more information about the terminology used in PM, see Grattan-Guinness
2000: 384–400.
22 For a succinct introduction to various non-classical logics, see Priest 2001.
23 For more biographical information about Hilbert, see Reid 1996, and, for more
information about Formalism in general, see Kreisel 1958, Detlefsen 1993, and
Hintikka 1995.
Notes to pages 39–65 203
24 As mentioned in section 2.4, Hilbert seems to have known about Cantor’s own doubts
concerning set theory as early as 1896. For more information, see Grattan-Guinness
2000: 117–119.
25 For instance, see Hilbert 1932[1918], in which he considers the utility of such a
language when constructing axiomatic-deductive arguments.
26 For timely words of caution see Ewald 1996: 1106–1107 and Mancosu 1998: 163–
164.
27 By far the best source of information concerning all aspects of Brouwer’s life and
work is van Stigt 1990.
28 It is of interest, incidentally, that Brouwer’s topological papers largely conform to
classical assumptions about mathematics and do not overtly manifest his Intuitionist
concerns. No doubt his reticence during these years was partly due to his desire to
ingratiate himself with his peers.
29 ‘Brouwer, that is the revolution!’
30 For more information concerning Intuitionistic logic see Mancosu 1998 (part 4) and
van Stigt 1990, especially chapter 5. For a discussion of the relationship between
Intuitionism and proof theory, see Kino et al. 1970.
31 For more information concerning Intuitionist set theory see van Stigt 1990
(chapter 6) and Mancosu 1998: 1–27.
Mathematical linguistics
1 These particular examples can be found in Bloomfield 1926: 154.
2 For more information concerning the influence of Weiss upon Bloomfield’s work,
see Belyi 1967. Bloomfield cites Weiss’ paper (Weiss 1925) explicitly in his own
paper (Bloomfield 1926).
3 Certain aspects of Bloomfield’s later work concerning this topic are discussed in
section 3.7.
4 These particular examples can be found in Bloch 1948: 6–7.
5 Another important example, namely Bar-Hillel’s development of Categorial Gram-
mar, is discussed at length in section 3.4.
6 It should be noted that Chomsky discussed Harwood’s paper explicitly in Syntactic
Structures and compared Harwood’s system with his own version of phrase struc-
ture grammar. For specifics, see Chomsky 1957b: 26n1. Some of the reasons why
the axiomatic-deductive character of TGG impressed contemporary linguists are
considered in various places in chapters 4 and 5.
7 Detailed discussions of the set of primitive recursive functions can be found in
Kleene 1952 (chapter 9), Crossley and Dummett 1965, and Fitting 1981. For an
approachable and insightful analysis of the various terminological problems that
have enveloped recursive function theory since the late 1930s, see Soare 1996.
8 The well-known schemata discussed here were influentially presented in Kleene
1952: 219.
9 For more details, see Gödel 1986b[1934] and Kleene 1952: 270–276.
10 Various aspects of the work of Bar-Hillel are discussed in several sections of this
chapter. This focus is necessary since his role in the development of TGG has often
204 Notes to pages 65–77
21 Quine’s own detailed account of his time in Vienna can be found in Quine
1985: 86ff. A useful recent assessment of Quine’s work is Orenstein 2002, which
includes some discussion of the complex relationship between Chomsky and
Quine.
22 The complete text of Quine’s 1934 lectures on Carnap can be found in Creath
1990: 47–103.
23 For specifics concerning Carnap’s interpretation of the connection between his work
and that of Goodman, see Carnap 1963: 19.
24 In brief, Goodman realised that ‘joints’ and ‘segments’ were more effective measures
of simplicity. An n-place predicate joins each occupant of one of its n places with
the occupants of the other places. Therefore, Goodman was able to define joints
as follows: ‘The number of joints . . . is the number of places in any minimum
set of pairs of places that ties all the places of the predicate together – i.e., that
provides between every two places in the predicate a route consisting entirely of
steps between places paired in the set’ (Goodman 1949: 36). By contrast, the notion
of a segment can be expressed in terms of place occupants that are identical, and the
guiding intuition is that an n-place predicate is more complex if fewer of its places
have the same occupant. As Goodman puts it: ‘since all the places that have the
same occupant as any given place belong to what I call a segment, clearly segments
rather than places are to be counted in determining complexity’ (Goodman 1949:
37).
25 An analytic truth is one that is true for logical reasons (e.g., p → ( p ∨ ¬ p)), while
a synthetic truth depends upon extralogical considerations (e.g., ‘Socrates is a man’
is true or false depending upon the state of the universe of discourse).
26 For example, see Fries 1954. A broad discussion of these issues can be found in
Matthews 1993: 118–122. The intensity of Bloomfield’s distaste for semantics has
been questioned from time to time. For some discussion of this position, see Murray
1994: 130–132.
27 The remaining fragments can be found in Bloomfield 1970b: 333–338.
28 This fact alone suggests some kind of familiarity with the work of Russell (perhaps
his accessible The Principles of Mathematics (1903)?), since Russell had been the
most assiduous paradox collector, and, as mentioned in section 2.5, the theory of
logical types had been designed primarily to obviate the kind of set-theoretical self-
reference that engendered the paradoxes.
29 A clear presentation of Grelling’s paradox can be found in Grelling 1936.
30 This paper is mentioned briefly in Bloomfield (1970a[1937]: 333).
31 The redefinition of the notion of a set was one common response to the work of Can-
tor, Whitehead, and Russell during the first half of the twentieth century. For instance,
as discussed in section 3.4, in the 1920s Leśniewski’s ‘mereology’ was devised in
order to avoid some of the problems of self-reference. For more information, see
Luschei 1962.
32 The task of attempting to determine which researchers unambiguously belonged
to the group now known as the post-Bloomfieldians has preoccupied a number
of linguistic historiographers. For a consideration of the various complexities, see
Hymes and Fought 1981[1975].
206 Notes to pages 109–123
8 It is worth mentioning that in his (now usually ignored) 1954 review of Rieger’s Mod-
ern Hebrew, which had been published in 1953, Chomsky commented favourably
upon the type of data collection techniques that Rieger advocates, remarking that
they constitute ‘a method that a linguist might be tempted to use in constructing a
linguistic corpus’ (Chomsky 1954: 180). This brief observation may indicate that,
in 1954, Chomsky was still primarily concerned with the sort of post-Bloomfieldian
corpus-based discovery procedures that he would later reject, or, at least, it may
reveal that he was not yet willing to question the validity of these methods in public.
9 For example, see the discussion in Katz 1981, chapter 1. Katz’s assessment of the
topic is a good example of the sort of opportunities for explication that have been
missed in the past. His approach is philosophical rather than historiographical, and
therefore he is too swift to project some of Chomsky’s later ideas (especially his
later ‘rationalism’) onto his earlier work. The title of Katz’s chapter, ‘From Harris’s
Nominalism to Chomsky’s Conceptualism’, indicates the nature of the deficiency.
Also, inexplicably, Katz does not mention Carnap in the chapter, despite his obvious
awareness of the influence of Carnap’s work on linguistics in the 1950s.
10 See Chomsky 1975[1955]: 294–306 for full details concerning various problems
with phrase structure grammar.
11 For example, see the succinct and (comparatively) non-technical discussion in Chom-
sky 1957b: 61–84.
12 For details, see Matthews, 1993: 133ff.
13 It should be noted that, although it appeared in the 1975 publication of LSLT, the
introduction (part of which is quoted here) is dated 1973.
14 For this particular example, see Chomsky 1975[1955]: 175.
Conclusion
1 Feyerabend is here quoting Lovis Rosenfeld, and he has added the italics himself.
2 i.e., Bloomfield 1926, Sapir 1929, Lees 1957, Hockett 1967, Chomsky 2000b.
3 For similar diatribes, see Postal 2004.
Bibliography
209
210 Bibliography
Clark, R. W. (1975), The Life of Bertrand Russell, London: Jonathan Cape and
Weidenfeld & Nicolson.
Collins, C. (2002), ‘Eliminating Labels’, in Epstein and Seely 2002: 42–64.
Copi, I. (1971), The Theory of Logical Types, London: Routledge & Kegan Paul.
Creath, R. (ed.), (1990), Dear Carnap, . . . , Dear Van: the Quine–Carnap
Correspondence and Related Work, Berkeley: University of California Press.
Crossley, J. N., and Dummett, M. A. E. (1965), Formal Systems and Recursive
Functions, Amsterdam: North-Holland.
Dauben, J. W. (1979), Georg Cantor: His Mathematics and Philosophy of the Infinite,
Cambridge, Massachusetts: Harvard University Press.
Dauben, J. W. (1995), Abraham Robinson: The Creation of Nonstandard Analysis: A
Personal and Mathematical Odyssey, Princeton: Princeton University Press.
Davis, S., and Mithun, M. (1979), Linguistics, Philosophy and Montague Grammar,
Austin: University of Texas Press.
Dedekind, R. J. W. (1932[1872]), Stetigkeit und irrationale Zahlen, reprinted in Fricke,
R., Noether, E., and Ore, O. (eds.), (1930–1932), Gesammelte mathematische
Werke, Vieweg: Braunschweig; III: 315–334.
Dedekind, R. J. W. (1932[1888]), ‘Was sind und was sollen die Zahlen?’, reprinted in
Fricke, R., Noether, E., and Ore, O. (eds.), (1930–1932), Gesammelte
mathematische Werke, Vieweg: Braunschweig; III: 335–391.
Detlefsen, M. (1993), ‘Hilbert’s Formalism’, Revue Internationale de Philosophie
47: 285–304.
Dugac, P. (1973), ‘Eléments d’analyse de Karl Weierstrass’, Archive for the History of
Exact Science 10: 41–176.
Dummett, M. A. A. (1977), Elements of Intuitionism, Oxford: Clarendon Press.
Dummett, M. E. (1991), Frege: Philosophy of Mathematics, London:
Duckworth/Cambridge, Massachusetts: Harvard University Press.
Embleton, S., Joseph, J. E., and Niederehe H.-J. (eds.) (1999), The Emergence of
Modern Language Science, 2 vols., Amsterdam: John Benjamins.
Epstein, S. D., and Hornstein, N. (eds.) (1999), Working Minimalism, Cambridge,
Massachusetts: MIT Press.
Epstein, S. D., and Seely, T. D. (eds.) (2002), Derivation and Explanation in the
Minimalist Program, New York and London: Blackwell.
Euler, L. (1912[1755]), Institutiones Calculi Differentialis, reprinted in Opera Omnia,
series 1, vol. 10: 69–72.
Ewald, W. (1996), From Kant to Hilbert: A Source Book in the Foundations of
Mathematics, vols. I and II, Oxford: Clarendon Press.
Feyerabend, P. (2001[1975]), Against Method, revised 3rd ed, London and New York:
Verso.
Fitting, M. (1981), Fundamentals of Generalised Recursive Function Theory,
Amsterdam: North-Holland.
Fought, J. (ed.) (1999), Leonard Bloomfield: Critical Assessments of Leading
Linguists, London and New York: Routledge.
Frampton, J., and Gutmann, S. (2002), ‘Crash-Proof Syntax’, in Epstein and Seely
2002: 90–105.
Bibliography 213
Hilbert, D., and Bernays, P. (1934–1939), Grundlagen der Mathematik, vols. I and II,
Berlin: Verlag von Julius Springer.
Hintikka, J. (1995), From Dedekind to Gödel: Essays on the Development of the
Foundations of Mathematics, London and New York: Kluwer.
Hiorth, F. (1974), Noam Chomsky: Linguistics and Philosophy, Oslo, Bergen, Tromsφ:
Universitetsforlaget.
Hjelmslev, L. (1961[1943]), Omkring sprogteoriens grundlaeggelse, reprinted in
English translation as Prolegomena to a Theory of Language by F. J. Whitfield,
Madison: University of Wisconsin Press.
Hockett, C. F. (1953), review of Shannon and Weaver’s The Mathematical Theory of
Communication, Language 29: 69–93.
Hockett, C. F. (1955), A Manual of Phonology, Chicago: University of Chicago
Press.
Hockett, C. F. (1967), Language, Mathematics and Linguistics, The Hague: Mouton de
Gruyter.
Hockett, C. F. (1968), The State of the Art, Janua Linguarum Series, Minor, 73, New
York: Cornell University Press.
Hrbacek, K., and Jech, T. (1984), Introduction to Set Theory, 2nd edn, New York and
Basel: Marcel Dekker, Inc.
Huck, G. J., and Goldsmith, J. A. (1995), Ideology and Linguistic Theory: Noam
Chomsky and the Deep Structure Debates, New York and London: Routledge.
Huybregts, R., and Riemsdijk, H. V. (1982), Noam Chomsky on the Generative
Enterprise, New Jersey: Foris Publishers.
Hylton, P. (1990), Russell, Idealism, and the Emergence of Analytic Philosophy,
Oxford: Oxford University Press.
Hymes, D., and Fought, J. (1981[1975]), American Structuralism, The Hague, Paris,
and New York: Mouton.
Jech, T. (1991), Set Theory, 2nd edn, Berlin and New York: Springer.
Johnson, P. E. (1972), A History of Set Theory, Boston: Prindle, Weber, & Schmidt.
Kasher, A. (ed.), (1991), The Chomskyan Turn, Oxford: Basil Blackwell.
Katz, J. (1981), Language and Other Abstract Objects, Oxford: Basil Blackwood.
Kino, A., Myhill, J., and Vesley, R. E. (eds.) (1970), Intuitionism and Proof-Theory:
Proceedings of the Summer Conference at Buffalo, N.Y., 1968, Amsterdam:
North-Holland.
Kleene, S. (1952), Introduction to Metamathematics, Amsterdam: North-Holland.
Kneale, W., and Kneale, M. (1962), The Development of Logic, Oxford: Clarendon
Press.
Koerner, E. K. (1999), Linguistic Historiography: Projects and Prospects, Amsterdam:
John Benjamins.
Kreisel, G. (1958), ‘Hilbert’s Programme’, Dialectica 12: 346–372.
Kuhn, T. (1962), The Structure of Scientific Revolutions, Chicago: University of
Chicago Press.
Lagrange, J. L. (1867[1760]), ‘Essai d’une nouvelle méthode pour déterminer les
maxima et les minima des formules intégrales indéfinies’, Miscellanea Taurinensia
2: 173–195, reprinted in Oeuvres, I (1867), 355–362.
216 Bibliography
Murray, O. S. (1994), Theory Groups and the Study of Language in North America: A
Social History, Amsterdam: John Benjamins.
Nagel, T. (1995) Other Minds: Critical Essays, 1969–1994, Oxford: Oxford University
Press.
Newmeyer, F. J. (1980), Linguistic Theory in America: The First Quarter-Century of
Transformational Grammar, New York and London: Academic Press.
Newmeyer, F. J. (1986), Linguistic Theory in America, 2nd edn, Orlando, Florida:
Academic Press.
Newmeyer, F. J. (1996), Generative Linguistics: A Historical Perspective, London:
Routledge.
Newton, I. (1967[1665]), ‘A Method for Finding Theorems Concerning Quaestiones
de Maximis et Minimis’, in Whiteside, D. T. (ed.), The Mathematical Papers of
Isaac Newton 1664–1666, Cambridge: Cambridge University Press, vol. I, 272–297.
Newton, I. (1969[1670]), ‘De methodis serierum et fluxionum’, in Whiteside, D. T.
(ed.), The Mathematical Papers of Isaac Newton 1664–1666, Cambridge:
Cambridge University Press, vol. III, 32–353.
Newton, I. (1972[1687/1726]), Philosophiae Naturalis Principia Mathematica, 3rd
edn, ed. A. Koyré and B. Cohen; 2 vols., Cambridge: Cambridge University Press.
Orenstein, A. (2002), W.V. Quine, Boston, Massachusetts: Acumen Press.
Nunes, J. (1999), ‘Linearization of Chains and Phonetic Realization of Chain Links’,
in Epstein and Hornstein 1999: 217–250.
Otero, C. P. (1994a), ‘The Emergence of Transformational Generative Grammar’, in
Otero 1994b, 1–36.
Otero, C. P. (ed.), (1994b), Noam Chomsky: Critical Assessments, 4 vols., New York
and London: Routledge.
Peano, G. (1959a[1889]), Arithmetices Principia Nova Methodo Exposita, reprinted in
Unione Matematica Italiana (eds.), Opera scelte 2: 20–55.
Peano, G. (1959b[1891]), ‘Principii di Logica Matematica’, reprinted in Unione
Matematica Italiana (eds.), Opera scelte 2: 92–101.
Post, E. (1944), ‘Recursively Enumerable Sets of Positive Integers and their Decision
Problems’, Bulletin of the American Mathematical Society 50: 284–316.
Postal, P. M. (2004), Skeptical Linguistic Essays, Oxford: Oxford University Press.
Priest, Graham (2001), An Introduction to Non-Classical Logic, Cambridge:
Cambridge University Press.
Quine, W. V. O. (1940), Mathematical Logic, Cambridge, Massachusetts: Harvard
University Press.
Quine, W. V. O. (1946), ‘Concatenation as a Basis for Arithmetic’, Journal of
Symbolic Logic 11: 105–114.
Quine, W. V. O. (1951), Mathematical Logic, revised 2nd edn, Cambridge,
Massachusetts: Harvard University Press.
Quine, W. V. O. (1953), From a Logical Point of View, Cambridge, Massachusetts:
Harvard University Press.
Quine, W. V. O. (1985), The Time of My Life: An Autobiography, Cambridge,
Massachusetts: MIT Press.
Reichenbach, H. (1947), Elements of Symbolic Logic, New York: Macmillan.
218 Bibliography
220
Index 221
Church, A., 15, 49–51, 55, 61, 64 Copenhagen Circle, 101, 107
classical sets, 69 copy theory of movement, 196
cognitive function, 190 corpora, 71, 114–116, 118, 120, 121, 124, 132,
competing grammars, 115, 116, 119, 120, 149, 150, 154, 155, 168, 169
154–156, 185, 194, 195 counterargument, 83
computability theory, 61, 130, 190–195 Curry, A., 96
computer science, 15, 199
Conant, J. B., 110 D’Alembert, 26, 27
concatenation, 78, 86, 117, 157–159 data, 2, 103, 154, 177
conjectures, 177 Dauben, J. W., 202
conjunction, 34, 36, 41, 42, 65, 86, 92 Davis, S., 207
consistency of axioms, 43 De Morgan, A., 33
consolidated grammar, 118, 119 declarative, 52
constant function, 64 Dedekind, R. J. W., 28, 29, 31, 46, 61
constructional system theory, 114 deduction, 3, 27, 35, 36, 55, 91, 102, 130, 132,
and Carnap, 74, 75, 79, 151, 204 152, 179, 186
and constructional levels, 204 deep-structure, 199
and discovery procedures, 124, 144 definienda, 122
and extralogical bases, 77, 80, 112, 117, definiendum, 37
118, 158 definition, 27, 41, 55, 63, 77, 91, 102, 117,
and extralogical primitives, 82 128, 129, 140, 156, 176–178
and formal languages, 14 dependence, 39
and Formalism, 122 derivation, 41, 43, 75, 83, 171, 174, 179, 189,
and Goodman and Quine, 85 196
and linguistic levels, 140, 158 derivatives, 23, 25
and logical analysis, 106 Descartes, 22
and multigrade relations, 80 descriptive adequacy, 193
and nominalism, 85, 87, 121 descriptivism, 17
and simplicity, 80–84, 113, 118, 121, 155, desideratum, 131
185, 193 differential equations, 201
and syntactic analysis, 121–123 differentiation, 22–24, 27, 28, 202
and syntactic theory, 123, 124, 158, discourse analysis, 161
159 discourse structure, 162
influence on syntactic theory, 73, 122 discovery procedures, 19, 71, 104, 124, 140,
influence on TGG, 9, 12, 14, 17, 19, 117, 149–151, 155, 169, 185, 208
125, 131, 140, 151, 154–156, 185, 197, discrete expression, 193
206 disjunction, 36, 41, 42, 76
neglect of in historiographical studies of distributional linguistics, 103, 127, 128, 176,
TGG, 14 183, 204
the development of, 19, 54 Dugac, P., 202
constructive nominalism, 7, 8, 12, 19, 85, Dummett, M. A. A., 202
87, 89, 109, 122, 156, 185, 186, 197,
198 economy, 14, 80–84, 113, 114, 116, 120, 167,
continuity, 47 195, 196, 198
continuous number line, 28 electromagnetism, 34
continuum, 46, 47 elegance, 196
contradictions, 32, 43, 69, 102 Elegant Syntax, 196
conversations, 85 elementary syntax, 49
conversions, 117–120, 160, 170–173, 181, elements, 3, 30, 31, 69, 75–77, 80, 92, 115,
196 119, 123, 127, 157, 160, 178, 190
224 Index
empiricism, 7–9, 15, 17–19, 65–67, 84, 87, and Formalism, 39, 44, 89–91, 102, 128, 203
123–125, 132, 151–153, 155, 156, 185, and Intuitionism, 45
207 and Leibniz, 32
epistemology, 74, 75, 89, 153, 207 and logic, 60, 69, 73, 78, 84, 88–90, 153
EPP feature, 197 and logical syntax, 49
EST, 188, 199 and Logicism, 33, 68
ethics, 74 and mathematics, 97
Euler, L., 24, 25, 27 and meaning, 128–129, 207
Europe, 44, 78, 102 and natural languages, 49, 52, 53, 58, 69, 70,
evaluation procedures, 12, 19, 115, 117, 125, 73, 90, 123, 127, 130, 139, 182, 184
140, 149, 150, 155, 156, 185, 195, 206 and nominalism, 69, 85, 86, 123
Ewald, W., 203 and science, 90
existential operator, 41 and simplicity, 119
explanatory adequacy, 193, 194 and syntactic theory, 56, 70, 73, 123, 124,
explanatory theory, 150 128, 131, 133, 135, 178, 187
extralogical bases, 118, 124, 158 and the axiomatic-deductive method, 56,
extralogical postulate, 81 59
extralogical primitives, 80–82, 153 formal linguistics, 102, 106, 121, 158,
175–177, 193
faculty of language, 190, 191, 194 formal notation, 160, 161
fallacy, 98, 100 formal proofs, 42
Fermat, P. de, 22 formal relations, 134
Feyerabend, P., 196 formal rules, 90
Fine, N., 110 formal sciences
finite alphabets, 178 and Formalism, 95
finite arithmetic, 30, 38 and linguistic theory, 1, 4, 7, 16, 18, 66, 71,
finite classes, 165 73, 95, 104, 109, 183
finite formulae, 100 and recursion, 64, 65, 191, 192
finite grammars, 168–171, 193 and TGG, 1, 2, 4, 5, 15, 18–19, 140, 159,
finite lexicon, 58, 59, 178 174, 187, 188
finite mathematics, 50, 60, 63, 169, 190, 193 and the axiomatic-deductive method, 56
finite proofs, 63, 168, 179 and the post-Bloomfieldians, 10, 183
finite sequences, 52, 64, 69 definition of, 2, 3
finite sets, 30, 31, 85, 142, 158, 169, 178, 190, in the twentieth century, 21, 54
193 formal symbols, 41, 43, 50
finite state grammars, 143, 145–148 formal syntax, 49, 53, 54, 86, 89, 91, 93–94,
finite state machines, 141, 144, 148 96, 105, 117, 128, 166, 180
finite strings, 123 formal systems, 14, 40, 43–45, 49, 50, 52, 58,
Fitch, T., 190–193 65, 73, 80, 89, 91, 95, 102, 128, 137, 178,
FLN hypothesis, 190–192 179, 186
fluxional calculus, 23, 201 and TGG, 5
formal consequence, 134, 162 formal theories, 5, 8, 53, 106, 107, 121, 155,
formal deduction, 89 160, 162, 176, 177, 208
formal discourse, 94 Formalism
formal features, 188, 197 and Bloomfield, 94–96, 101, 104, 183
formal grammars, 119–121, 125, 169, 179, and Carnap, 90, 91, 181
190, 193, 195 and Chomsky, 122
formal inference, 43 and Church, 49, 50
formal languages and formal languages, 52
and concatenation, 78, 158 and Gödel, 45
Index 225
Intuitionism, 18, 21, 32, 45, 46, 51 and logical empiricism, 9, 151
irrational numbers, 100 and mathematics, 101
iteration, 171 and meaning, 13, 95, 132, 136, 138, 180
and recursion, 65, 66, 170
Jakobson, R., 143 and simplicity, 113, 115, 116, 119, 196
Jech, T., 202 and stochastic processes, 141, 143, 144
Jena, 73 and syntax and semantics, 18, 179
Joos, M., 102 and TGG, 1–3, 7, 9, 11, 19, 108, 110, 156
and the axiomatic-deductive method, 2, 57,
Kasher, A., 204 178
Katz, J., 208 and the formal sciences, 1, 4–7, 54, 183, 200
kernel, 165, 173, 174 and the foundations crisis, 96
Kleene, S., 15, 50, 51, 54, 55, 61, 63, 65, 203 and the post-Bloomfieldians, 6, 10, 110, 176
Kneale, W., 202 and transformations, 127
Koerner, E., 17, 201 linguistics, 183–186
Köhler, 76 and constructional system theory, 19, 122
Kuhn, T., 16, 110 and empirical science, 67
and Formalism, 5, 60, 89, 101, 102, 107, 176
labels, 190, 196 and historiography, 13
Lagrange, J. L., 26 and logic, 5, 89, 94, 105–107, 125,
Landen, J., 24 130–134, 136–138, 159, 208
language faculty, 194 and mathematics, 5, 6, 19, 54, 57, 93, 94,
language of science, 97 96, 103, 175
Latin, 90 and nominalism, 122, 155
Lees, R., 1–3, 5, 188, 198, 208 and philosophy, 8
Leibniz, 22, 24, 32, 201 and psychology, 95
Lenneberg, E., 187 and recursion, 65
Leonard, 80 and semiotics, 102
Leśniewski, S., 67–70, 85, 103 and simplicity criteria, 115
lexical features, 188 and statistics, 143, 148
lexical items, 189, 190 and structuralism, 102
lexical properties, 189, 197 and the axiomatic-deductive method, 27,
Lexical-Functional Grammar, 12 55–57, 60, 177
lexicalist theory, 197 and the formal sciences, 16, 21, 32, 48, 73,
lexicon, 197 95, 105, 106, 198
limits, 23, 26, 28, 100, 147 and the foundations crisis, 97, 98
linear grammar, 120, 171 and the post-Bloomfieldians, 151, 176
linguistic levels, 115, 117, 140, 156, 157 at MIT, 16
linguistic theory Chomsky’s introduction to, 110
and constructional system theory, 14, 151, distributional linguistics, 104, 110, 127
155, 158 structural linguistics, 127
and discovery procedures, 150, 154, 185 Linsky, L., 133
and distribution procedures, 128 logic
and evaluation procedures, 154 and algebra, 33
and explanatory adequacy, 193 and analysis, 26, 27
and Formalism, 102, 104, 121, 177 and arithmetic, 39
and linguistic levels, 157, 158 and constructional system theory, 79
and logic, 12, 19, 51, 52, 73, 94, 109, 121, and Formalism, 39, 49–51, 89, 175, 182
124, 131, 133, 135–137, 158, 162, 176, and inference, 162
184 and Intuitionism, 46, 203
228 Index
metamathematics, 10, 13–15, 43, 44, 48–51, and meaning, 69, 87, 132, 134, 141, 181,
65, 89, 102, 106, 175, 176, 178, 179, 181, 182
182 and Montague Grammar, 139
metaphysics, 74, 78 and pragmatics, 69
Miller, G. A., 143, 187 and recursion, 67, 169, 170, 174
minimal search, 197 and simplicity, 113, 195, 206
Minsky, M., 110 and statistical modelling, 11, 141–149
MIT, 16, 145, 175, 206 and structuralism, 143
models, 142, 145 and synonyms, 133
modern science, 199 and TGG, 1, 141, 177
Montague, R., 138, 139, 207 and the axiomatic-deductive method, 2, 5,
Montague Semantics, 138 55, 57, 58, 184
morphemes, 103, 112, 123, 127, 144, 157, and the foundations crisis, 98, 135, 136
204 and the foundations of mathematics, 103
MP, 194, 195, 197, 198 and the real world, 2
multivariate calculus, 21 and transformations, 161–164
Murray, O. S., 12, 16, 17, 205, 206 and Zipfian distributions, 141
complexity of, 134, 156
Nagel, T., 207 English, 52, 53, 139
natural languages French, 65
adequacy criteria for, 112, 130, 193 German, 90, 207
and and constructional system theory, 123 interdisciplinary approaches to the study of,
and Categorial Grammar, 70, 72 9, 10, 175
and cognition, 56 Latin, 90
and competing grammars, 154 Polish, 207
and constructional system theory, 14, 121, properties of, 194
122 natural numbers, 64
and constructive nominalism, 89 natural philosophy, 23
and Esperanto, 53 natural science, 199
and finite-state grammars, 145, 147 negation, 36, 37, 41, 42, 76
and formal consequence, 134 neologisms, 181
and formal languages, 52, 53, 58, 70, 90, 95, network, 83, 116
123, 134–136, 139, 147, 178, 182 Neurath, O., 73, 93
and Formalism, 183 Newmeyer, F. J., 11–13, 16, 207
and grammaticality, 60, 120 Newton, 22–24, 201
and induction, 152, 153 nominalism, 7, 8, 54, 68, 69, 79, 82, 85–89,
and inference, 134 121–124, 155, 158, 201, 208
and infinity, 168, 169 non-standard analysis, 202
and knowledge, 175 North America, 16
and linguistic data, 154 noun phrase, 65, 117
and linguistic levels, 157 number theory, 21, 26–32, 38, 39, 43, 48, 51,
and linguistic theory, 2 202
and linguists, 127, 133 number-theoretic functions, 61–63
and logic, 38, 51, 52, 58, 59, 73, 89, 94, 97, numeration, 189
126, 130, 131, 137, 158, 159, 175, 176,
184 object language, 86
and logical epistemology, 74 ontology, 68, 75
and logical syntax, 90, 92, 93, 104 optimality conditions, 120
and mathematics, 45, 56, 94, 97, 101, 104, ordinal numbers, 31
105, 142, 183 ordinary language, 175
230 Index
Orenstein, A., 205 principles, 5, 9, 14, 22, 33, 35, 40, 42, 46, 47,
Otero, C. P., 15, 206 151, 158, 193, 199
Oxford, 9, 175 Principles and Parameters, 12, 188, 195, 199,
206
paradigm, 150 probability, 33, 142, 144, 149
paradoxes, 21, 32, 35, 69, 85, 99, 100, 205 proof theory, 40–45, 48–51, 60, 62, 65, 89,
Pascal, 22 102, 128, 160, 168, 174, 178–180, 184
passive transformation, 166 proofs, 27, 41–43, 45, 49, 179, 183, 186
passives, 166 propositional calculus, 33, 36, 49
Peano, G., 28, 33, 34, 61, 62, 65 propositional functions, 37, 52
Peirce, C., 33 psychology, 6, 7, 11, 56, 58, 67, 76, 92, 95,
permutation, 165, 166 112, 130
philosophy, 2, 6, 8, 15, 33, 34, 45, 46, 67, 68, pure mathematics, 2
73, 74, 84, 87, 88, 105, 110, 136,
151–155, 175, 207 qualities, 78, 113, 153
philosophy of language, 133 quantification, 33, 48, 87, 118
philosophy of mathematics, 104 quantifiers, 49
phonemes, 57, 119, 127, 141, 143, 164, 204 quantum logic, 38
phonemic analysis, 57, 58, 143, 157, 171 quantum mechanics, 15
phonetic analysis, 143, 158 quasi-analysis, 76
phonetic features, 98, 99 Quine, W. V. O.
phonological analysis, 143, 194 and analytic and synthetic truths, 87, 88, 133
phonological form, 188 and concatenation, 78, 86, 158
phrasal projections, 194, 197 and constructional system theory, 121, 140,
phrase, 98 197
phrase structure grammar, 120, 145, 157, 164, and constructive nominalism, 85–88, 122
166, 167, 170, 172–174, 178, 181, 197, and Goodman, 78, 80, 84
203, 208 and logic, 48, 49, 55
phrase structure level, 172 and logical empiricism, 9, 18, 135, 151, 153
physics, 199 and logical epistemology, 87
Piattelli-Palmarini, M., 199 and logical syntax, 78
Platonism, 85, 86, 197 and meaning, 135
poetry, 141 and simplicity, 80, 81, 116–118, 197
Polish logicians, 67, 71, 102 and the Vienna Circle, 77
Post, 15, 54, 61, 64, 104, 169, 170 his autobiography, 205
post-Bloomfieldians, 6, 7, 9, 10, 12, 13, 15, 16, his Ph.D., 77
18, 71, 102–104, 107, 136, 141, 143, 146, influence of Carnap, 74, 77–79, 85, 89
149–152, 154, 168, 169, 175, 176, 183, influence on Chomsky, 7, 8, 12, 14, 18, 88,
205, 208 110, 112, 118, 150, 153, 154, 159, 185,
Postal, P., 198, 199 186
postulates, 55–58, 116, 130, 133, 184 influence on Harris, 105
Prague, 77 neglect of in historiographical studies, 8
Prall, D., 78 relationship with Chomsky, 110, 111
predicate calculus, 36, 37, 49, 50
predicates, 62, 80, 82, 83, 86, 118, 123 rational numbers, 28
predictive grammars, 169 rationalism, 7, 9, 17, 88, 154, 195, 207
primitive recursion, 63 real economy, 118
primitive recursive functions, 63, 64, 203 real numbers, 39, 47
primitives, 75, 80, 81, 83, 91, 116, 122, 133, recursion, 19, 43, 61–63, 65, 66, 96, 167, 168,
188, 190, 197 171–173, 188, 190–193
Index 231
Recursion Convention, 192 sentence, 52, 53, 58, 60, 65, 66, 70–72, 87, 88,
recursive components, 60, 64, 169–174, 190, 91–94, 117, 128, 129, 145–148, 159, 160,
197 164, 166, 173, 178, 180, 186
recursive definitions, 19, 62–67, 106, 131, 137, sentence generation, 179
140, 168, 170, 184, 186, 191, 193 sentential calculus, 68, 103
recursive function theory, 10, 43, 50, 54, set theory, 18, 21, 29–37, 39, 40, 46–48, 50,
60–65, 67, 73, 96, 104, 168, 169, 174, 51, 60, 68–70, 86, 96, 98, 100, 101, 122,
175, 191–193 202, 203, 205
recursive rules, 12, 19, 170–174 sets, 30–32, 34, 47, 64, 69, 100, 123, 134, 157,
recursive steps, 169, 189 165, 168, 169, 197
recursively enumerable languages, 174 Shannon, C., 141–146, 148
recursively enumerable sets, 61, 64, 168–170 Sheynin, O. B., 141
reductionism, 157 similarity relations, 76, 124
Reichenbach, H., 51, 52, 105 simplicity, 185
relativity theory, 15 and constructional system theory, 9, 12,
Richardson, A., 204 80–84, 116, 118, 193
Rieger, 208 and linguistic theory, 9, 12, 113, 115–117,
Riemann, G. F. B., 29 130, 196
rigorisation, 18, 29, 55, 67 and recursion, 197
rigour, 2, 7, 26–28, 55, 64, 107, 137 and TGG, 109, 113–120, 151, 155, 156,
Rosenbloom, P. C., 52 166, 171, 193–195
Rosenfield, 208 simplicity criteria, 2, 14, 18, 19, 59, 81, 82,
Russell, B., 34–39, 42, 48, 49, 60, 67, 69, 70, 87, 106, 112–121, 125, 149, 155, 156,
76, 98, 111, 122, 205 166, 167, 171, 172, 185, 193–195, 197,
Russell’s paradox, 68, 98 198, 206
Russia, 109 theory-external, 195
theory-internal, 81, 84, 113, 121, 132,
Sapir, E., 103, 104, 198, 208 195–197, 206
Saussure, 102 Skolem, 103
Schaffer, S., 201 Smith, L., 207
schemata, 64, 203 Smith, N. V., 206
scheme, 71, 85, 153 Soare, R. I., 192
Schlick, M., 74, 77, 93 sociology, 130
Schröder, E., 33 solipsism, 46
Schnitt, 28 spatial intuition, 38
Schützenberger, M., 187 Spec–Head, 197
science, 6, 40, 57, 65, 74, 89, 90, 94, 97, 104, speech forms, 97, 100
130, 153, 185, 192, 198–200 spell-out, 196
science of language, 55 statistical modelling, 3, 11, 141, 143–145, 148,
scientific language, 97 149
scientific method, 3, 186 statisticians, 141
scientific theory, 2, 9, 16, 90, 116, 151, 158 Steinberg, D., 17
self-reference, 35, 41, 62, 96, 98, 99, 190, 193, Stevenson, C., 78
205 stochastic grammars, 140, 141, 143–145, 148,
semantic, 70 149
semantics, 3, 18, 19, 68, 70, 95, 96, 102, 106, stochastic processes, 19, 141–143, 145, 148
121, 126, 128–130, 134–136, 138, 141, structural linguistics, 11, 18, 71, 72, 95, 111,
149, 179–181, 183, 186, 194, 205 125, 127, 130, 138, 143
semiotics, 102 structuralism, 10
sense data, 76, 153 Struik, D. J., 201
232 Index
subject-predicate analysis, 33, 52 Tarski, A., 67, 78, 85, 105, 129, 130, 204,
subsets, 30, 32, 58–60, 90, 165, 176 207
substrings, 181 tautology, 36, 88, 129
successor function, 64 Tawardowski, K., 67, 68
Suppe, F., 201 Taylor, B., 24, 27
surface-structure, 199 terminal symbols, 164
syllables, 141 TGG
symbols, 45, 49, 50, 56, 72, 91, 92, 95, 103, and Bar-Hillel, 184, 203
106, 123, 142, 145, 160, 164, 165, 176, and constructional system theory, 14, 73,
179 125
synonymy, 88, 127, 133, 134, 136 and constructive nominalism, 8, 12, 197
syntactic categories, 164 and discovery procedures, 71, 124
syntactic classes, 124 and Goodman, 185
syntactic objects, 189, 190, 196 and induction, 152
syntactic theory, 197 and logical empiricism, 155, 159
and constructional system theory, 73, and logical syntax, 73, 105
112–114, 121, 123–125, 140 and meaning, 91, 181
and constructive nominalism, 122, 123, 186 and Montague Grammar, 139
and discovery procedures, 154 and previous syntactic theories, 5
and empiricism, 17, 156 and recursion, 60, 64, 67, 140, 168–170,
and explanatory adequacy, 193 172–175, 190
and Formalism, 44, 122, 136, 175, 177, and simplicity, 14, 59, 84, 117, 120, 121,
179–182, 184, 186 185, 193–195, 206
and induction, 13 and the axiomatic-deductive method, 5, 60,
and logic, 8, 69, 70, 73, 89, 91, 109, 114, 125, 177, 178, 203
184 and the formal sciences, 2, 15, 19, 96, 174,
and logical empiricism, 77 176, 186, 187
and logical epistemology, 74 and the history of ideas, 6
and mathematics, 10, 11, 14, 21, 22, 58, 105 and the post-Bloomfieldians, 7, 10–13, 16,
and nominalism, 19, 86 18, 169
and philosophy, 15 and transformations, 91, 93, 140, 159,
and recursion, 12, 60, 61, 64, 66, 67, 168, 163–168
169, 184 as a scientific theory, 1, 4, 16, 198, 199
and simplicity, 117, 132 as an explanatory theory, 150
and stochastic processes, 142 Chomsky on the origins of, 8, 10, 151
and TGG, 4, 6, 10, 11, 16, 108, 111, 114, definition of, 3
140, 149 Goodman’s knowledge of, 111
and the axiomatic-deductive method, 58, 60, Harris’ attitude towards, 111
184, 188 Hockett on the origins of, 6
and the formal sciences, 1, 4, 16, 19, 109, in the 1960s, 13, 14
199 influence of analytic philosophy, 88
and the post-Bloomfieldians, 10, 12 influence of Carnap, 104, 186
and transformations, 92, 140, 164 influence of Formalism, 50, 140, 174, 179,
categorial grammar, 70, 72, 126 180, 186, 188
generative grammar, 188 influence of logic, 8
syntacticians, 131, 141, 168 influence of Logicism, 60
syntax, 70, 90, 91, 95, 126, 129, 132, 138, 140, influence of Quine, 78, 88
141, 143–145, 149, 161, 163, 169, 179, influence of the formal sciences, 4, 5, 18,
180, 186, 191, 207, 208 19, 21
synthetic truth, 205 influence of White, 186
Index 233
other histories of, 2, 4–8, 12–17, 186, 187 UG, 4, 195, 206
reception of, 2 ungrammaticality, 59, 60, 66, 148
the definition of, 4, 108 universal operator, 41
the development of, 1, 4, 17, 21, 54, 102,
107, 108, 140, 183, 186–188, 195, Van Stigt, W. P., 46, 203
198 verb phrase, 117
the reception of, 1 verbs, 56, 65, 143
theorems, 2, 3, 5, 26, 27, 31, 32, 45, 50, 63, 81, verifiability, 49, 180
103, 123, 124, 178 Vienna, 67, 73, 74, 77, 205
theoretical syntax, 49, 50 Vienna Circle, 67, 73, 74, 77, 89, 93, 94, 98,
theory construction, 1 204
theory of descriptions, 163
theory of functions, 25, 27 Warsaw, 68, 207
theory of logical types, 35, 36, 68, 70, 122, Warsaw-Lvov school, 130
202, 205 Weaver, 141–144, 148, 207
theory of meaning, 135, 180 Weierstrass, K., 28, 46, 202
theory of reference, 135 Weiss, A. P., 56, 203
theory of relations, 76 well-formedness, 70, 179, 180
theory of simplicity, 84 well-ordering, 100
theory-internal economy, 113, 114, 195 Wells, R., 102–105, 175
theory-internal simplicity, 19, 81, 84, 121, Wertheimer, 76
132, 193–197, 206 Weyl, H., 46
thought and language, 45 wh-questions, 166
traditional grammars, 107 White, M., 87, 88, 110, 133, 135
transfinite arithmetic, 31, 32, 38, 42 Whitehead, A. N., 15, 34–39, 42, 60, 67–69,
transfinite set theory, 29 76, 77, 111, 122, 202, 205
transformational level, 157, 164 Wilder, R. L., 51, 55
transformations, 12, 19, 90–93, 102, 119–121, Wittgenstein, 111
126, 127, 133, 140, 145, 159–168, 173, Wolénski, J., 204
174, 181, 186, 207 word formation, 53
translation, 89, 126, 137, 168, 176, 204, 207
trigonometric functions, 29 X-bar theory, 194, 197, 199
truth-conditional semantics, 2, 128, 129, 139,
207 yes-or-no questions, 166
Turing, A., 61, 104, 191, 192 Young, J. W., 48, 51, 55
Turing computable functions, 192
Turing machine, 191, 192 Zermelo, E., 31
Turing Thesis, 192 Zipf, G., 141