You are on page 1of 12

THREE MODELS FOR THE DESCRIPTION OF LANGUAGE


Noam Chomsq
Department of Modern Languages and Research Laboratory of Electronics
Massachusetts Institute of Technology
Cambridge, Massachusetts
Abstract
We investigate several conceptions of observations, to show how they are interrelated,
linguistic structure to determine whether or and to predict an indefinite number of new
not they can provide simple and II reYealing" phenomena. A mathematical theor,y has the
grammars that generate all of the sentences additional property that predictions follow
of English and only these. We find that no rigorously from the body of theor,y. Similarly,
finite-state Markov process that produces a grammar ia based on a finite number of observed
symbols with transition from state to state sentences (the linguist's corpus) and it
can serve as an English grammar. Furthermore, "projects• this set to an infinite set of
the particular subclass of such processes that grammatical sentences·by establishing general
produce n-order statistical approximations to 11 laws11 (grammatical rules) framed in terms of
English do not come closer, with increasing n, such eypothetical constructs as the particular
to matching the output of an English grammar. phonemes, words, phrases, and so on, of the
We formalize.. the notions of 0 phra.se structure" language under analysis. A properly formulated
and show that this gives us a method for grammar should detel"llline unambiguously the set
describing language which is essentially more of grammatical sentences.
powerful, thoue;h still representable as a rather
elementary type of finite-state process, Never­ General linguistic theory can be viewed as
theless, it is successful only when limited to a a metatheor,y which is concerned with the problem
small subset of simple sentences. We study the of how to choose such a grammar in the case of
formal properties of a set of grammatical trans­ each particular language on the basis of a finite
formations that carry sentences with phre.se corpus of sentences. In particular, it will
structure into new sentences with derived phrase consider and attempt to explicate the relation
structure, showing that transformational gr!l.111mars between the set of grammatical sentences and the
are processes of the same elementary type as set of observed sentences. In other words,
phrase-structure grammars; that the grammar of linguistic theory attempts to explain the ability
English is materially simplified if phrase of a speaker to produce and understand new
structure description is limited to a kernel of sentences, and to reject as ungrammatical other
simple sentences from which all other sentences new sequences, on the be.sis of his limited
are constructed by repeated transformations; and linguistic experience.
that this view of linguistic structure gives a
certain insight into the use and understanding Suppose that for many languages there are
:ef language. certain clear cases of grammatical sentences and
certain clear cases of ungrammatical sequences,
l. Introduction e.g., (l) and (2), respectively, in English.
There are two central problems in the ( l) John ate a sandwich
descriptive studT of language. One primary (2) Sandwich a ate John.
concern of the linguist ia to discover simple
and 11 revealingll grammars for natural languages. In this case, we can teat the adequacy of a
At the same time, by studTing the properties of proposed linguistic theor,y by determining, for
auch auccessful gramna.rs and clarifying the basic each language, whether or not the clear casea
conceptions that underlie them, he hopes to are handled properly by the grammars constructed
arrive at a general theol'1' of linguistic in accordance with this theor,y. For example, if
structure. We shall examine certain features of a large corpus of English does not happen to
theae related inquiries. contain either (1) or (2), we ask whether the
grammar that ifl detern,ined for this corpus will
The grammar of a language can be viewed as project the corpus to include (l) and exclude (2i
a theory of the structure of this language. Any Even though such clear cases may provide only a
scientific theor,y is based on a certain finite weak test of adequacy for the grammar of a given
set of observations and, by eatablishing general language taken in isolation, they provide a very
laws stated in terms of certain eypotbetical strong test for any general linguistic theory and
constructs, it attempts to account for these for the set of grammars to which it leads, since
we insist that in the case of each language the
clear cases be handled properly in a fixed and
*This work was supported in part by the Army predetermined manner. We can take certain steps
(Signal Corps), the Air Force (Office of Scientific towards the construction of an operational
Research, Air Research and Development Command), characterization of "grammatical sentence" that
and the Navy ( Office of Naval Research), and in will provide us with the clear cases required to
part by a grant from Eastman Kodak Company. set the task of linguistics significantly.

113
Observe, for example, that (l) will be read by an of linguistic structure in terms of the possi­
English speaker with the normal intonation of a bility and complexity of description (questions
sentence of the corpt1s, while (2) will be read (J) , ( 4))• Then, in § 6, we shall briefly
with a falling intonation on each word, as will consi�er the same theories in te?'J!ls of (5), and
any sequence of unrelated words. Other dis­ shall see that we are independently led to the
tinguishing criteria of the same sort can be same conclusions as to relative adequacy for the
described. purposes of linguistics.
Before we can hope to provide a satisfactory 21 Finite State Markov Processes.
account of the general relation between observed
sentences and grammatical sentences, we must 2.1 The most elementary gramme.rs which, with
learn a great deal more about the formal proper­ a finite amount of apparatus, will generate an
ties of each of these sets. This paper is con­ infinite number of sentences, are those based on
cerned with the formal structure of the set of a familiar conception of language as a
grammatical sentences. We shall limit ourselves particularly simple type of information source,
to English, and shall assume intuitive knowle dge namely, a finite-state Markov process.l Spe­
of English sentences and nonsentences. We then cifically, we define a finite-lilli grammar G
ask what sort of linguistic theory is required as as a system with 1} finite number of states
a basis for an English grammar that will des cribe S 0,.. s
, , a set A=Laijkl O.$ i,j.$q; 1Sk .$N ij for
4
t
the set of English sentences in an interestir.g each i,j}or transition symbols, and a set
and satisfactory manner. C= { (Si,Sj of certain pairs of states of G that
The first step in the linguistic analy"sis of a.re said to be connected. As the system moves
a language is to provide a finite system of from states to S j' it produces a symbol aijktA.
representation for its sentences. We shall assume 1
Suppose that
that this step has been carried out, and we
shall deal with languages only" in phonemic (6) s ,••,so:.
or alphabetic transcription. By a language, °1. m
then, we shall mean a set (finite or infinite) is a sequence of states of G with �=-a..-o, an:i
of sentences, each of finite length, all (S ,S ),c for each i<m. As the system mV88
constructed from a finite alphabet of symbols. a.1 o.i+l
If A is an alphabet, we shall say that anything ftom s to s it produces the symbol
formed by concatenating the symbols of A is a 0:.1 a.1+1
string in A. By a grammar of the language Lwe
mean a device of some sort that produces all of (7)
the strings that are sentences of Land only
these. A
for some k <N a. • .Using the arch to signify
No matter how we ultimately" decide to - ui i+l
construct linguistic theory, we shall surely" concatenation, 2 we say that the sequence (6)
require that the grammar of any language must be generates all sentences
finite. It follows that only" a countable set of
grammars is made available by arry linguistic ( 8)
theory; hence that uncountably" many languages ,
in our general aense, are l1 terally not describable
in terms of the conception of linguistic structure for all e,ppropriate choices of k 1 (i.e., for
provided by any particular �lieory. Given a ki.$ N ) • The language L conte.ining all and
o:.io:.i+l G
proposed theory of linguistic structure, then, it
is always appropriate to ask the following question: onl,y such sentences is called the language
generated by G.
(J) Are there interesting languages that a.re
simply" outside the range of description of the Thus, to produce a sentence of L we set the
proposed type? system G in the initial states0 Gand we run
through a sequence of connected states, ending
In particular, we shell ask whether English is a.gain withs , and producing one of the associated
such a language. If it is, then the proposed transition sjmbols of A with each transition from
conception of linguistic structure must be judged one stets to the next. We say that a language L
ine.d.equate. If the answer to (3) is negative, we is a .f!ai1!-ste.te language if Lis the set of
go on to ask such �u.estions as the following: sentences genera.ted by some fini te-sta.te grammar G.
(4) Can we construct reasonably" simple 2.2. Suppose that we take the set A of transition
gramme.rs for all interesting languages? symbols to be the set of English phonemes. wa
(5) Are such gramme.rs II revealing" in the can attempt to construct a finite state grammar G
sense that the syntactic structure that they which will generate every string of English
exhibit can support semantic analysis, can provide phonemes which is a grammatical sentence of
insight into the use and understanding of language, English, and onl,y such strings. It is illlJllediatel,y
etc.? evid.ent that tho task of constructing a finite­
state grammar for Ep.glish can be considerablY
We shall first examine various conceptions simplified if we take A as the set of English

114
morphemeeJ or words, and construc t Geo that it example. the languages L,,½• L., described in
will generate exa.ctq the grammatical s tringsof (1 2) are not describable'oy any finite-state
these uni ts. we can then complete the grammar grammar. ,.... ,.... ,....
b7 giving a finite set of rules that give the (12) (1) L contains a b, a a0 b b,
� l"\
phonemic spelling of each word or morpheme in l"'\ l"\
a a a" b b b•••• , and 1n
each con text in which it occurs. We shall general. all sentences consisting
consider briefq the status of such rules in of n occurrences of a followed by
§ 4.1 and § 5.J. exactq n occurrences of b, and only
these; ,... I"'\ I"'\ I"\ I"\
Before inquiring direc tq into the problem (ii) L2 contains a a, b b. a b b a,
of cone t ru.cting a finite-state gra.mmar for
English morpheme or word sequences, le t us b l"\ a f'\ a f'\ b, a "a f'\ bf'\ b f'\ af'\ a•••••
investiga te the absolute limi ts'of t he set of and in general, all 11 mirror-image 11
finite-s tate languages. S1Jl>PO•e tha t A is the sentences consisting of a s tring X
alphabet of a language L, t ha t a1 ,.••an are followed by X in reverse. and only
I""\ t hese; f'\ " f'\ f'\ "
s71Dbols of thil alphabet, and tha t S=a1I""\•• • an (111) � contain• a a, b b, a b a b,
is a sentence of L. We sq that S has an (i,j)­ f'\
b af'\ b a, a "' af'\ b " a"al"'\ b,••• ,
dependencz with respect to L if and onq if tile and in general, all sentences con­
following condi tions are met: sisting of a s tring X followed by
1 $1<j$n the iden tical s tring x. and only
(9)(1) t hese.
(ii) ·there are symbols b1t b_1 €A with the In½• for example, for any m we can find a
property that S1 is not a sentence of sentence with a dependency set.
L. and s2 is a sentence of L. wher e S1 4
Dm= { (1,2m) ,(2,2m-l),••• (m,m+l)} •
is formed from S by replacing the 1th:
symbol of S (namely, ai) by bi. and Sz 2.3. Turning now to English. we find that there
is formed from S1 by replacing t he are infinite sets of sentences that have depend.8JX1'
Jth symbol of S1 (namely, aj) by bj • sets with more than any fixed number of terms.
For example. let S 1,S2,••• be declarative sent8l'Xl88,
In other words, S has an (1,j)-depenitncy with Then the following are all English sentences:
respect to L if replacement of the i symbol a ( 13)(1) If s1 1 then Sz.
i
of S by b (b jai) requires a corresponding (11) Either SJ, or S4 .
i 1 (111) The man who said that S5, is
replacement of the J th symbol a j of s by b arriving today.
j
(bJ/aJ) for the resulting s tring to belong to L. These sentences have dependencies between •1f"-
11th en", •eith er"-•or•, •man•-•1s•. But we can
We sq that »•{ (°1_,a1),••• (a.m•�m)}1s a choose s1 , s3, S5 which appear between the inter­
dependent words, as (lJi). (1311), or (13111) them­
dependency s..!! for s in L if and onq 1f the selves. Proceeding to construct sentences in this
following conditions are met: wq we arrive at subparts of English with j ust the
mirror image propert ies of the languages L1 and L2
(10)(1) For·lS1Sm. S has an (a.1.�i )­ of (12). Consequently, English fails condition
dependency with respect to L (11). English is not a finite-st ate language. ana:·
(11) for each i,j, we are forced to rej ect the theory of language
a.1< aJ under discussion as failing condition (J).
(111) for each i,j such that ilJ, a. �J We might avoid this consequence by an
1 arbitrary decree that there ii a finite upper
and a 1,aJ. limi t to sentence leng th in English . This would
Thus, in a dependency se t for S in L ever.r two serve no useful purpose, however. The poin t is
dependencies a:r:e distinct 1n both terms and each that there are processes of sentence formation
"determining" element in S precedes all 11 de­ that this elemen tary model for language is
termined• elements, where we picture a as intrinsicalq incapable of handling. If no
determining the choice of a . a. 1 finUe limit is set for the operation of these
ai processes, we can prove the literal inapplica­
Evident]T, if S has an m-termed dependency s et bility of this model. If the processes have a
in L, at least 2• states are necessar.r in the limit , then the construction of a finite-state
fini te-state grammar that generates the grammar will not be li terally impossible (since
language L. a lis t is a trivial finite-sta t e grammar), but
t his grammar will be so complex as to be of li t tle
Thia observation enables us to s tate a use or interest. Below, we shall studl' a model
necessar.r condition for fini te-state languages. for grammars t hat can handle mirror-image lan­
guages, The extra power of such a model in the
(11) Suppose that L is a finite-sta te infinite case is reflec t ed in the fac t that i t is
language. Then there is an m such that no much more useful and revealing if an upper limi t
sen tence S of L baa a dependency set of more is set. In general, the assumption that langu8iJS
tbA.n m tel'llls in L. are infinite is made for the purpose of simpli­
With this condi tion in mind, we can easily fying the description.5 If a grammar has no
construct many nonfinite-s ta te languages. For recursive steps (closed loops, in the model

u,
discussed above) it will be prohibitively complex- - that there is apparently no significant approach
it will, in fact, turn out to be little better to the problems of grammar in this direction.
than a list of strings or of morpheme class
sequences in the case of natural languages. If it Notice that although for every n, a process
does have recursive devices, it will produce of n-order a�pro�imation can be represented as a
infinitely many sentences. finite-state Markov process, the converse is not
true. For example, consider the three-state
2.4 Although we have found that no finite-state process with (S0 ,s 1), (s 1 ,s 1),(s 1 ,s 0),
Markov process that produces sentences from left
to right can serve as an English grammar, we (S 0,S ) ,(s2,s2),(s2,S0) as its only connected
2
might inquire into the possibility of construe� states, and with a, b, a, c, b, c as the respect­
a sequence of such devices that, m some nontrivial ive transition symbols. This process can be
w,q, come closer and closer to-matching the output represented by the following state diagram:
of a satisfactory English grammar. Suppose, for
example, that for fixed n we construct a finite­
state grammar in the following manner: one state C a
of the grammar is associated with each sequence of
English words of length n and the probability that (16) b b
the word X will be produced when the system ia in
the state Si is equal to the conditional proba­
bility of X, given the sequence of n words which C a
defines S • The output of auch grammar is
customari!y called an n+1st order approximation to
Engliah. Evidently, as n increases, the output of Thia process can)>roduce the sentences a"a,
such grammars will come to look more and more like a ,... b"a, a ,... b"b a, a"b'"' b"b"a, ••• , c"c,
English, since longer and longer sequences have a c ,... b ,... c, c"b"b"c, c"b"b"b"" c, ••• , but not
high probability of being taken direct ly from the a "' b"b "' c, c ,... b"b "' a, etc. The generated
sample of English in which the probabilities were language has sentences with dependencies of any
determined. This fact has occasionally led to finite length.
the suggestion that a theory of linguistic
structure might be fashioned on such a model. In § 2 . 4 ve argu.ed that there is no
significant correlation between order of approxi­
Whatever the other interest of statistical m�tion and grammaticalness. If we order the
approximation in this sense m,q be, it is clear strings of a given length in terms of order of
that it can shed no light on the problems of approximation to English, we shall find both
grammar. There is no general relation between the grammatical and ungrammatical strings scattered
frequency of a string (or its component parts) and throughout the list, from top to bottom. Hence
its grammaticalness. We can see this most clearly the notion of statistical approximation appears
by considering such strings as to be irrelevant to grammar. In§ 2.J we pointed
out that a much broader class of processes,
(14) colorless green ideas sleep furious ly neJDely, all finite-state Markov processes that
produce transition symbols, does not include an
which is a grammatical sentence, even though it is English grammar. That is, if we construct a
fair to assume that no pair of its words m,q ever finite-state grammar that produces only English
have occurred together in the past. Notice that a sentences, we know that it will fail to produce
speaker of English will read (14) with the an infinite number of these sentences; in par­
ordinary intonation pattern of an English sentence, ticular, it will fail to produce an infinite
while he will read the equally unfamiliar string xrwnber of true sentences, false sentences,
reasonable q_ue-ations that could be intelligibly
( 15) furiously sleep ideas green colorless asked, and the like. Below, we shall investigate
s still broader class of processes that might
with a falling intonation on each word, as in provide us with an English grammar.
the case of any ungrammatical string. Thus (14)
differs from (15) exactly as ( 1) differs from (2); 3. Phrase Structure.
our tentative operational criterion for gram­
maticalness supports our intuitive feeling that J.l. Customariq, syntactic description is
( 14) is a grammatical sentence and that ( 15) is given in terms of what is called "immediate
not. We might state the problem of grammar, in constituent analysis." In description of this
part, as that of explaining and reconstructing sort the words of a sentence are grouped into
the ability of an English speaker to recognize phrases, the.se are grouped into smaller consti­
( 1), (14), etc., as grammatical, while rejecting tuent phrases and so on, ·until the ultimate
(2) , (-15), etc. But no order of approximation constituents (generally morphemes3) are reached.
model can distinguish ( 14) from (15) (or an These phrases are then classified as noun
indefinite number of similar pairs). As n phrases (NP), verb phrases (VP), etc. For
increases, an nth order approximation to English example, the sentence (l?) might be ana ly zed as
will eJCcluile ( as more and more imt>robableJ an in the accompanying diagram.
ever-1ncreas1ng number of grammatical sentences
while it still contains vast numbers of complet; ly
ungrammatical strings.6 We are forced to conclude

ll.6
(17) the man took the book that are derivable from aome [ I ,J'] grammar,
and we sq that Lis a terminal language if it is
BP Verb BP the aet of terminal atringa from some a7atem
( I ,F].
VP
Sentence In every interesting case there will be a
terminal vocabulary VT (VT C VP) that
exactq characterizes the terminal strings, in
Evidentq, description of sentences in such t81'111a the sense that every terminal string is a string
permits considerable simplification over the in VT and no s7mbol of VT is rewritten in aey of
word-b7-vord model, aince the composition of a the rules of J'. In such a case we can interpret
complex class of expreaaions such as BP �an be the terminal strings as constituting the language
stated just once in the grammar, and this class under anaqsis (with V as its vocabular,r), and
can be uaed as a building block at various the derivation■ of theAe strings as providing
points in the construction of aentencea. We now their phrase structure.
ask what form of grammar corresponds to thi•
conception of linguistic structure. J.J. AB a simple example of a s7stem of the form
(18), consider the following small part of EngUsh
J.2. A phrase-stl'llcture grammar ia defined b7 a grammar:
finite vocabulary (alphabet) VP, a finite eet I
(20) I : #"sentence
,....
of initial strings in V , and A finite set Fof #"'
rulee of the form: X -P.Y, where X and Y are Sentence - NP vp
F:
strings in VP. Each such rule is interpreted aa VP - Verb"' NP
the instruct�on: rewrite X as Y. J'or reasons NP - the ,... man, the"' book
that will appear directq, we require that in Verb - took
each such [ I , J'] gramu.r Among the derivations from (20) we have, in
particular:
(18)
(21) D : #"sentence"#
l ,..
# BP"VP"# ""'
# ,..BP"Verb"BP #
#"the" man,....Verb" lllP"#
#"the" man"Verb,.... the" book "#
#"the" man " took" the" book" #
D2 : #"Sentence"#
Yi is formed from X by the replacement of a # ,...BP"VP"#
i
single symbol of x1 by som<3 string. Neither #"the" man"VP"#
#"the" man '"'Verb'"' lllP"#
the replaced symbol nor the replacing string #"the" man '"' took"BP" #
may be the identity element U of footnote 4. #"the" man" took" the" bookr"I#
Given the ( I ,J'] grammar (18), we sq that: fheae derivations are evidentq equivalent; the7
differ onq in the order in which the rules are
(19)( 1) a string �,.{ollovs !'.l:2!!l a string a. applied. We can represent this equivalence
if a.=Z "x1 W and �"Y 1,.... w, for graphicalq b7 constructing diagrams that
some 1 Sm;? correspond, in an obvious wq, to derivations.
(ii) a derivation of the string S is a
Both D1 and n2 reduce to the diagram:
sequence D=(s1 , •• ,st) of strings, #"sentence l"\ #

/�
(22)
where s 1c I and foreach 1< t, S i+l
follows from s1;
BP
( 111) a string S 1a derivable from ( 18)
if there ia a derivation of S in /\_
the man Verb
/'-....BP
terms of (18);
(iv) a derivation of S t is terminated if I /"'-..
took the book
there ia no string that followsf1'0III The diagram (22) gives the phrase structure of
St ; the terminal sentence •the man took the book,"
a string st 1a a terminal string if just as in (l?). In general, given a derivation
it is the last line of a terminated D of a string S, we BST that a substring s of S
derivation. is an X if in the diagram corresponding to D, s
is traceable back to a single node, and this node
A derivation is thu.s roughq analogous toa is labelled X. Thu.a given D1 or D2l, correspond­
proof, Yi th I taken as the axiom 17stem e.nd F ing to (22), we say that 11 the " me.ni is e.n NP,
11 took the" book" is a VP, 11 the book11 is an
"' "'
aB the rulea of inference. We •BT that Lis a
derivable language if Lis the set of strings NP, "tbe "me.n" took " the " book11 is a Sentence.
"me.n" took,11· however, is not a phrase of this

ll.7
string at all, Rince it is not traceable back such rules. In J J.2 we required that in such
to an,v node. a rule as (25), X must be a single symbol. This
ensures that a phrase-structure diagram will be
When we attempt to construct the simplest constructible from any derivation. The �ramma.r
possible [ I ,F] grammar for English we find that can also be simplified very greatly if we order
certain sentences automatically receive non­ the rules and require that they be applied in
equivalent derivations, Along with (20), tbe sequence (beginning again with the first rule after
grammar of English will certainly have to c<ntain applying the final rule of the sequence), and if
such rules as we distinguish between obligatorz rules which
(2J) Verb-are "' flying must be applied when we reach them in the sequence
Verb-are and optional rules which ma;r or ma7 not be
BP-they applied. These revisions do not modify the
BP-planes generative power of the grammar, although they
BP-flying '"' planea lead to considerable simplification.
in order to account for such sentences as •they It seems ·reasonable to require for
are flying - a plane• (1iP-Verb-1iP), "(flying) aignificance aome guarantee that the gramu.r will
planes - are - noisy• (NP-Verb-Adjective), et.c. actually generate a la- rge number of sentence■ in
But this set of rules provides us with two :aon­ a limited 8IIIOWlt of time; more apecifically, that
equivalent derivations of the sentence •they are it be imposaible to run through the sequence of
flying planes•, reducing to the diagrams: rule■ vacuously (applying no rule) unless the
last line of the derivation under construction is
(24) #"sentence"# a te rminal atring. We can meet thia requirement

✓ "w
#�te�#
by poaing certain condition■ on the occurrence of
BP VP obligato17 rules in the sequence of rules, We
I
they Verb
/'-....BP I
they
/ "'-..1iP
Ver'b
define a proper grammar aa a 17atem [I.�].
where I is a set of initial strings and� a
I /"'-.
are flying plane■
/" I
are flying planes
sequence of rules x 1- Y1 as in (18), with the
additional condition that for ea�h i there must
be at least one j such that X 1=Xj and_ X 1_ :-Y 1_ ia
Hence this sentence will have two phrase an obligatory rule, Thu.a, each.- Ieft-haM te1'11 of
structures assigned to it; it can be 1U1alyzed as the rulea of (18) muit appear in at least one
•they - are - fl.,ing planes• or •they - are f'.cy-ing obligato17 rule, Thia ia the weakest aimple
- planes.• And in fact, this sentence ia condition that guarantee■ that a nonterminated
ambiguous in just this wa;r; we can underatand it derivation must advance at least one step eve17
as meaning that •those specks on the horizon - time we run through the rulea. It provide• that
are - flying planes• or "those pilots - are flying if x1 can be rewritten as one of yi ,••,Y1
- planes.• When the simplest grammar automatic­ l k
ally provides nonequivalent derivation■ tor some then at least one of these rewritings must take
sentence, we sa;r that we have a case of place, However, :e.roper grammars are esaentially
constructional homowmitz, and we can suggest different from [ l: ,F] grammar■, r.t D(G) be
this formal property as an explanation for the the aet of derivations producible from a phrase
atrupture grammar G, whether pro�r or no� r.t
semantic ambiguity of the sentence in question.
In �l we posed the requirement that grammars D •l D(G)
F
I I
G a [ I ,FJ grammar J 8.lld D�=1_1)( G) G
offer insight into the use and understanding of a proper grammar} • Then
language (cf. (5))• One wa;r to teat the adequac1
of a grammar is by determining whether or not (26). D and D are incomparable; i.e.,
F Q,
the cases of constructional homo�ity are D
actually cases of semantic ambiguity, as in ( 24) � ¢ � and D� ¢ �•
We return to this important problem in § 6,
That is, there are systems of phrase structure
In (20)-(24) the element # indicated that can be deacribed by [ I ,F] grammars but not
sentence (later, word) boundary, It can be by proper grammar■, and. others that can be
taken as an element of the terminal vocabulary described by proper grammars but not by [ I ,F]
v discu11ed in the final paragraph of f J,2. grammar■•
,_.
3,5, We have defined three types of language:
J,4, Theae segments of English grammar are much finite-state languages (in §2.1), derivable and
oversimplified in several respects. For one terminal languages (in §3.2). These are related
thing, each rule of (20) arid (23) has only a in the following way:
single symbol on the left, although we placed no
such limitation on [ I,F] grammars in § J.2 . (27)(1) every finite-state language 1s a
A rule of tbe form terminal language, but not conversely;
(ii) every derivable language is a terminal
(25) z "' x "' w-z"y"w language, but not conversely;
(111) there are derivable, nonfinite-state
indicates that X can be rewritten as Y only in languages and finite-state, nonderivable
the context Z--W. It can easily be shown that languages,
the grammar will be much simplified if we permit Suppose that LG is a finite-state language

ll8
with the finite-state gr!lllllllar G as in § 2.1. The major import of Theorem (27) is that
We construct a [ I ,:r] grammar in the following description in terms of phrase structure is
manner: I = { S 0j ; F conta.ina a. rule of the foI'lll euentialq more powerful (not just sb:,pler) than
(281) for each i,j,k such that (S 1,sj )cC, j-JO, description in terms of the finite-state grammars
and k S Bij; :r contains a rule of the form (i81i) that produce sentences from left to right. In
§ 2.J we found that English is li tera.lly beyond
for each i,k such that (s ,s )cC and k � Bio" the bounds of these grammars because of mirror­
1 0 in;age properties that it shares with L1 and 12
(28)(1) s1 - aijk"'S j of (12). We have just seen, however, that 1
(ii) Si- aiok 1
is a terminal language and the same ia true of
12· Renee, the considerations that led us to
Clearq, the ter11inal language from this [ I ,F] reject the finite-etate model do not similarly
grammar will be exactq L ' establishing the lead us to reject the more powerful phrase­
first part of (271). G structure model.
In § 2,2 we found that 1i_, 12
and 1:, of Note that the latter is more abstract than
( 12) were not finite-state langusges. 11 and 12, the finite-state model in the sense that symbols
that are not included in the vocabulary of a
however, are terminal languages. For 11, e.g., language enter into the description of this
we have the [ I ,:!!'] grammar language, In the terms of § J.2, VP proper:cy
includes VT. Thus in the ca.se of (29), we
describe 1 in terms of an elementZ which is not
1
in�; and in .the case of (20)-(24), we introduce
such symbols as Sentence, NP, VP, etc., which are
Thia establishes (271). not words of English, into the description of
English structure.
Suppose that 14 ia a derivable language with 3.6. We can interpret a [ I ,F] grammar of the
the vocabulary V = { a ,•. ,a}. Suppose that we form ( 18) as e. rather elementa:cy finite-state
add to the grammlr of 1L a finite aet of rules process in the following ws:y. Consider a system
4 that hes a finite number of states S , •• ,S .
a - b , where the bi 1 s are not in V P and are 0 q
ail di!tinct. Then this new grammar gives a When in atate S0, it can produce any ct the
terminal language which is simply a notational strings of I , thereby moving into a new state.
variant of 14• Thus every derivable language It• state at an;y point is determined. by the sub­
is also terminal. set of elements of x1 ,••,Xm contained as sub-
A• an example of a terminal, nonderivable strings in the last produced string, and 1t moves
language cohsider the language 1 containing just to a new state by applying one of the rules to
the strings 5
this string, thus producing anew string. The
system returns to state S with the production
(30) a"b, c" a b d, c"c a b d d,
"' "' "' "' "" "'
"' b" d "d "d, •••
of a terminal string. Th�s system thus produces
c "c" c"a derivations, in the sense of § '.3.2. The process
is determined at any point by its present state
An infinite derivable language must contain an and by the last string that has been produced,
infinite set of strings that can be arranged in a and there is a-finite upper bound on the amount
sequence s1,s , ••• in such a wq that for sana of inspection of this string that is necessary
2
rule X -Y, s follows from s _1 by application before the process can continue, producing a new
1 1 string tbat differs in one of a finite number of
of this rule, for each i >l. And Y in this rule ways from its last output.
must be formed from X by repll!cement of a single
symbol of X by a string (cf. (18)). This is
evident:cy impossible in the case of 1 . This It is not difficult to construct languages
language is, however, the termir,al laiguage given that are beyond the r ange of description of
by the following grammar: [ I ,F] greJlllllars. In fact, the language 1 of
3
(12111) is evidently not a terminal language. I
(31) I z do not know whether English is actua.l:cy a terminal
F z-a b
'"'
language or whether there are other actual
Z -c"Z"d languages that are literalfy beyond the bounds of
An example of a finite-sta.te, nonderivable phrase structure description. Renee I see no ws:y
language is the language 1 containing all am. to disqualify this theory of linguistic structure
only the strings consisting6 of 2n or Jn on the basis of consideration (J). When we turn
to the question of the complexity of description
occurrences of a, for n=l,2,•••• Language 11 (cf. (4)), however, we find that there are ample
of (12) is a derivable, nonfinite-state language, grounds for the conclusion that this theory of
with the initial string a"b and the rule: linguistic structure is fundamentally inadequate.
a"b -a"a '"' b" b, We shall now investigate a few of the problems

ll9
that arise when we attempt to extend (20) to a
full-scale grallllllar of English. ,Applying this rule to each of the three J.f" v
sequences in the last line of (33), we derive
4. Inadequacies of Phrase-Structure Grammar (35) # "'the"man" have" past." #"be "en " # "
take "ing "' #"thel"\ book"#.
4.1. In (20) we considered only one w,q of
developing theelement Verb, namely, as •took•. In the first paragraph of § 2.2 we mentioned
But even with the verb stem fixed there are that a grammar will contain a set of rules (called
a great many other forms that could appear in morphophonemic rules) which convert strings of
the context •the man -- the book,• e.g., •takes,• morphemes into atrings of phonemes. In the
"has taken," •has been taking," 1 is taking,• morphophonemice of English, we shall have such
•has been taken," •will be taking,• and so on. rules ae the following (we use conventional,
A direct description of this set of elements rather than phonemic orthograpey):
would be fairly complex, because of the heavy
dependencies among them (e.g. , •has taken• but (J6) have"past - had
not "has taking," "is being taken" but not •1s be "' en - been
being taking,• etc.). We can, in fact, give a take" ing - taking
ve:ey simple analysis of "Verb" as a sequence of will"past - would
independent elements, but only by selecting as can "past - could
elements certain discontinuous strings. For M ,... present - M
example, in the phrase "has been taking" we can walk"past - walked
separate out the discontinuous elements "has •• en," take"past - took
•be •• ing," and "take•, and we can then say that etc.
these elemer.ts combine freely. Following this
course systematically, we replace the last rule Applying the morphophonemic rules to (35) we
in (20) by derive the sentence:
()2) (i) Verb -Auxilia:ey"V (37) the man had been taking the book.
(ii) V - take, eat, •••
(iii) Auxilia:ey-C(M)(have "en) (be ,... ixlg) Similarly, with one major exception to be
(be" en) discussed below (and several minor ones that we
(iv) M-wili, can, shall, m,q, must ehall overlook here), the rules (32), (34) will
(v) C -pe.st, present give all the other forms of the verb in
declarative sentences, and only these forme.
The notations in (32111) are to be inter­
preted .as follows: in developing "Auxilia:ey• This ve:ey simple analysis, however, goes
in a derivation we must choose the unpe.renthe­ beyond the bounds of [ I ,F] grammars in several
sized element C, and we mq choose zero or more respects. The rule (34), although it is quite
of the pe.renthesized elements, in the order simple, cannot be incorporated within a [ I ,F]
given. Thus, in continuing the derivation grammar, which has no place for discontinuous
D1 of (21} below line five, we might proceecl elements. Furthermore, to apply the rule (J4)
all' :follows: to the last line of (33) we must know that " take"
is a V, hence, a v. In other words, in order to
(JJ)#"the"man"Verb" the" book "# appl( this rule it ie necessa:ey to inspect more
[from »1 of (21)] than just the string to which the rule applies;
#" tbe"man" Auxilia:ey" v"the" book I"\# it is necessa:ey to know some of the constituent
((321)] structure of thie etring, or equivalently
(cf.§ 3.3), to inspect certain earlier lines in
#"the"man"Auxiliary" take"the"book"# its derivation. Since (34) requires knowledge of
[(3211)] the 'history of derivat1on 1 of a string, it violates
# "the"man" c" have" en" be" ing"take" the elementary property of [ I ,F] grammars
the" book " # discuued in § 3.6.
((32111), choosing the elements C,
have"en, and be '"' ing] 4.2. The fact that this simple analysis of the
verb phrase as a sequence of independent� chosen
#" the" man" past" have" en"be "ing"take "' units goes beyond the bounds of.( I ,F] grammars,
the" book " # suggests that such grammars are too limited to
[(32v)] give a true picture of linguistic etructure.
Further study of the verb phrase lends additional
Suppose that we define the class Af as containing support to thie conclusion. There is one major
the affixes •en•, "ing", and the C's; and the limitation on the independence of the elements
cle.ss ., al including all V I s, M I s, "have• , and •be .11 introduced in (J2). If we choose an intransitive
We can then convert the last lir1e of (33) into a verb (e.g., 11 come,11 11 occur,11 etc.) es V in (32),
properly ordered sequence of morphemes by the we cannot select be "'en as an auxilia:ey. We can­
following ru le: not h�ve such phrases as "John has been come,11
"John is occuri-ed," and the like. Furthermore,
(34) J.f"v -v"Af " # the element be "' en cannot be chosen independently
of the context of the phrase 11 Verb. n If we have

120
the element "Verb" in the context II the man -- the 5. Transformational Grammar.
food," we are constrained not to select be "' en in ,5 .• l. Each grammatical tranaforma tion T will
applying (32), although we are free to choose aey essentially be a ru.le that converts every sentence
other element of (J2). That is, we can have •·the with a given conatituent structure into a new
man is eating the food," •the man would have been sentence with derived conatituent atru.cture. The
eating the food," etc., but not "the ma.n is eaten tranaform and ita derived structure muat be relat.ed
the food," "the man would have been eaten the food,11 in a fixed and conatant wq to the structure of
etc. On the other hand, if the context of the the transformed string, for each T. We can
phrase 11 Verb11 is, e. g� , 11 the food - by the man," characterize T by atating, in atru.ctural terma,
we are required to select be "' en. We can have the domain of strings to which it applies and the
• the food is eaten by the man, 11 but not "the food change that it effects on any such string.
is eating by the man,• etc. In• short, we find that Let ua suppose in the following diacussion
the element be '"' en enters into a detailed network that we have a [ l: ,F] grammar with a vocabulary
of restrictions which distinguish it from all the V and a terminal vocabulary VT C V ' ae in � 3.2.
P P
other elements introduced in the analysis 0£ "Verb"
in (32). Thia complex and unique behavior of In § 3.3 we showed that a [ I ,F] grammar
be "' en suggests that it would be· desirable to permit, the derivation of terminal 1tring1, and we
exclude it from (32) and to introduce passives into pointed out that in general a given terminal stril\e;
the grammar in some other wq. will have several equivalent derivations. Two
derivation, were aaid to be equivalent if they
There is, in fact, a very simple wq to reduce to the same diagram of the form (22), etc.9
incorporate sentences with be"en (i.e., paHives) Suppoae that D1,..,D n conatitute a maximal set
into the grammar. Notice that for every active of equivalent derivations of a terminal string
sentence such as •the man ate the food" we have a S. Then we define a phraae marker of S as the set
corresponding passive 1 the food was eaten by the of strings that occur as lines in the derivations
man" and conversely. Suppose then that we drop the
element be"en from (32111), and then add to the D1, •• ,D n. A string will have more than one phraae
�t'!I.IU!ar the following rule: marker if and only if it ha• nonequivalent
derivation, (cf. (24)).

(38 ) If S is a sentence of the form NP 1- Suppoae that K ia a phraae marker of S. We


Auxiliary-V-NP2, then the corresponding string 89¥ that
of the form NPz-Au:idliary"be "' en-V-by"' NP i is
also a sentence. (J9) (S,X) is analyzable into (Xi,•••X) if
and only if there are strings s1,••,a au.ch ?tiia.t
For example, if "the man - past - eat the (1) S=s "' ••• "' an n
food" (NP1-Auxiliary-V-NP ) is a sentence, then l
2 (11) for each iSn, X contains the string
•the food - past be en - eat - by the man" (NPz­
1 •• 8 1-1 X{' 8 1+1 • • 8 n
8
"' "' "' "'
Auxiliary "' be "'en-V-b7 "'NP1 ) is also a s_entence.
Rules (J4) and (36) would convert the first of
these into "the man ate the food" and the (40) In th1a case, 1 is an x in S with
reapect to K. 10 1 1
second into 11 the food was eaten by the man.•
The advantages of this analysis Of passives The relation defined in (40) is exactly the
are unmistakable. Since the element be"en has relation "is a• defined in§ 3.J: i.e., s1 is an
been dropped from (32) it is no longer necessary to X in the sense of (40) if and·only if 1 11 a
qualify (32) with the complex of restriction• a-abetring of S which is traceable back tA a aingle
discusaed above. ·. The fact that be "' en can occur node of the diagram of the form (22), etc., and
only with transitive verbs, that it is excluded in this node is labelled x •
the context •the man -- the food" and that it is 1
required in the eontext "the food -- by the man," The notion of analyzability defined above
is now, in each case, an automatic consequence of allowa us to specify precisely the domain of
the analysis we have just given. application of aey tranafo:rmation. We associate
A rule of the form (3 8), however, is well with each transformation a restricting class R
beyond the limits of phrase-structure grammars. defined as follows:
Like (34), it rearranges the elements of the string
to which it applies, and it requires considerable (41) Ris a re1tr1eting c� if and only if
information about the constituent structure of this for some r,m, R is the aet of sequence,:
string. When we carry the detailed atudy of English l l
syntax further, we find that there are maey other
x
1•·. , xr
cases in which the grammar can be simplified if the
[ l: ,F] system is supplemented by rules of the same
general form as (38). Let us call each such rule a
grammatical transformation. As our third model for
the deeeription of linguiatic structure, we now
consider briefly the formal propertiea of a trans­ where xf is a string in the vocabulary V P' for each
formational grammar that can be adjoined to the i,j. We then say that a string S with the phrase
[ I ,r] grammar of phrase atru.cture. 8 marker X belongs to the domain of the transformatl.on

121
T if the restricting class R associated with T tp(Y1 ,Y ;YJ;Y ,Y ) •
2 J 4 ,::J
contains a sequence (iq,...x;> into which (S,X) is t (Y1 , ••,Y ;Y ) = by Y1
a�alyzable. The domain of a trapsformation is p 4 4
t.lllls a set of ordered :pairs (S,KJ of a string S t (Y1, •• ,Yn ;Yn , •• ,Yr) = Y n for all nSr,4.
and a phrase marker I of s. A transformation mq p
be applicable to s with one phrase marker, but not The derived transformation t*
with a second phrase marker, in the case of a string ing effect: p thus has the follow-
s with ambiguous constituent structure. (48)(1) t*( - Y2"be"en - Y - bf'y
p Y1,•• ,Y4) = Y1 J 1
In particular, the paseive transformation (11) t;(the ,... man,past, eat, the,... food)•
described in (J8) baa associated with it a the"food - past"' be" en - eat - b1"the"
restricting class R containing j ust one sequence: man.
p
(42) R= { (BP, Au.xiliacy, V, BP)} • The rules (J4),(J6) carcy the right-hand aide of
P (4811) into "the food was eaten b7 the man,• Just
Thie transformation can thus be applied to l1J:IT as the1 carcy (4J) into the corresponding active
string that is analyzable into an NP followed by "the lll8ll ate the food.•
The pai r (R,t ) as in (42),(47) completely
an Auxilie.cy followed by e. V followed b7 an NP. charactsrizea thi pisaive transformation as
For exa.mple, it can be applied to the string (4J) described in (J8). R tells us to which strings
analyzed into substrings s1, .. ,a4 in accordance this tranafol'ltation a�pliea (given the phrase
with the dashes. markers of these strings) and how to subdivide
these strings in order to apply the transformation.
(4J) the man - :past - eat - the food. and t tells us what structural change to effect
on thi subdivided string.
5.2. In this way, we can describe in structural
tems the set of strings (with phrase markers) to A grammatical transfol'lll&tion is specified
which e.n;r t ransforiration applies. We must now completely by a restricting class Rand an
specif7 the structural change that a transformation elementacy transformation t, each of which is
effects on l1llT string in its domain. An elementarz finitely characterizable, as in the case of the
transformation t is defined b7 the following passive. It is not difficult to define rigorousl1
property: the manner of this specification, along the lines
aketched above. To complete the development of
(44) fo r each pair of integers n,r (n Sr ), transformational grammar it is necessary to show
there is a unique sequence of integers (a0 ,a1 , .. ,�) how a transformation automatically assigns a
derived phrase ma rker to each transform and to
and a unique sequence of strings in V (z ,•••Zit.+ ) generalize to traneformationa on sets of strings.
P 1 1
auch that (i)a0=0; k�O;lSajS r for 1 SJ�k;Y0=U11 (These and related topiqs are treated in reference
(J].) A transformation will then carcy a string S
(11) for each Y1,••,Yr ' with a ph rase marker X (or a set of such pairs)
into a string S 1 with a derived phrase marker X 1•
o
Zi
t(Y1•••,Yn; Yn•••, Yr)=Ya,... ,... y ,... Z"
81.
2 "
ya ••"y8k
2
" z.
-k+l:
S,J. From these considere.t1ons we are led to a
Tlms t can be understood as converting the picture of grammars as possessing a tripe.rtite
occurrence of Y in the context structure. Corresponding to the phrase structu re
n
analysis we have a sequence of rules of the form
( 45) Y l ,... ••,.. yn- 1,... __,... y n+l,... •• ,... y r
X-Y, e.g., (20), (2J), (J2). Following this we
have a seguence of transformational rules such as
. ,... (J4) and lJ8). Finally, we have a sequence of
into a certain string Y a z 1"•·"ya. " z k+l
morphophonemic rules such as (J6). 8€ain of the
o form X-Y. To generate a sentence from such a
.It
which ia unique, given the sequence of terms grammar we construct an extended derivation
(Y , •• ,Y ) into which Y1 " •." Yr is subdivided. beginning with an initial string of ,.. the phrase
t !arriei the string Y1-••-Yr into a new string structure grammar, e.g., #"Sentence #, as in
W1-•. -Wr which is related in a fixed way to (20). We then run through the rules of phrase
y -•• -Yr. More precisely, we associate with t the structure, producing a terminal string. We then
dlrived transfonation t•: apply certain transformations, giving a string of
morphemes in the correct order, perhaps quite a
(46) t• is the derived transfonation of t if different string from the original terminal string.
and only if for all Y1, ••,Yr ' t•(Y1,••,Yr)=W{' • •"'!-, Application of the morphophonemic rules converts
where W =t(Y1,••,Yn;Yn,••,Y) this into a string of phonemes. We might :nm
n r for each n:Sr. through the phrase structure grammar several times
We now associate with each transformation T and then apply a generalized transformation to the
an elementacy transformation t. For example, with resulting set of terminal strings.
the passive transformation (J8) we associate the In § J.4 we noted that it is advantageous to
elementacy transformation tp defined e.s follows: order the rules of phrase structure into a
sequence, and to distinguish obligatocy from
(47) optional rules. The same is true of the trans-
f orma.ti onal part of the grRIDDlar. In § 4 we dis­
cussed the transformation ('.34), which converts a

122
sequence affix-verb into the sequence verb-affix, in terms of such essentially formal criteria as
and the passive transformation (38). Notice that simplicity. In § 1 we suggested that there are
(34) must be applied in every extended derivation, other relevant considerations of adequacy for
or the result will not be a grammatical sentence. such theories. We can ask (cf,(5)) whether or
Rule (31.i), then, is an obligatory transformation. not the syntactic structure revealed by these
The passive transformation, however, may or may theories provides insight into the use and under­
not be applted; either way we have a sentence. The standing of language. We can barely touch on
passive is thus an optional transformation. This this problem here, but even this brief discussion
distinction between optional and obligatory trans­ will suggest that this criterion provides the
formations leads us to distinguish between two same order of relative adequacy for the three
classes of sentences of the language. We have, on models we have considered.
the one hand, a kernel of basic sentences that are
derived from thetendnal strings of the phrase­ If the grammar of a languege is to provide
structure grammar by application of only insight into the way the language is understood,
obligatory transformations. We then have a set of it must be true, in particular, that if a sentence
derived sentences that are generated by applying is ambiguous (understood in more than one way),
optional transformations to the strings underlying then this sentence is provided with alternative
kernel sentences. analyses by the grammar. In other words, if a
When we actually carry out a detailed study certain sentence Sis ambiguous, we can test the
of English structure, we find that the gral!llllar can adequacy of a given linguistic theory by asking
be greatly simplified if we limit the kernel to a whether or not the simplest grammar constructible
very small set of simple, active, declarative in terms of this theory for the langu,!ee in
sentences (in fact, proba.bly a finite set) such as question automatically provides distinct ways of
"the man ate the food,11 etc. We then derive generating the sentences. It is instructive to
questions, passives, sentences with conjunction, compare the Markov process, phrase-structure, and
sentences with compound noun phrases (e.g., transformational models in the light of this test.
"proving that tmorem was difficult," with the NP
"proving that theorem").12 etc., by transformatiom. In § J.J we pointed out that the simplest
Since the result of a transformation is a sentence [ I ,F] gramma.r for English happens to provide
with derived constituent structure, transformati<Jlll nonequivalent derivations for the sentence "they
can be compounded, and we can form questions from are flying planes," which is, in fact, ambiguous.
passives (e.g., "was the food eaten by the man"), This reasoning does not appear to carry over for
etc. The actual sentences of real life are finite-state grammars, however. That is there
usually not kernel sentences, but rather is no obvious motivation for assigning t;o
complicated transforms of these. We find, however, different paths to this ambiguous sentence in any
that the trensformations are, by and large, meaning­ finite-state grammar that might be proposed for
preserving, so that we can view the kernel a part of English. Such examples of construction­
sentences underlying a given sentence as being, in al homonymity (there are •any others) constitute
some sense, the elementary "content elements" in independent evidence for the superiority of the
terms of which the actual transform is l\mderstoed." phrase-structure model over finite-state grammars.
We discuss this problem briefly in§ 6, more
extensively in references [l], (2]. Fu.rther investigation of English brings to
In § J.6 we pointed out that a grammar of light examples that are not easily explained in
phrase structure is a rather elementary type of terms of phrase structure. Consider the phrase
finite-state process that is determined at each
point by its present state and a bounded amount of (49) the shooting of the hunters.
its last output. We discovered in§ 4 that this
limitation is too severe, and that the grammar can
be simplified by adding transforma.tional rules We can understand this phrase with "hunters" as
tha.t take into account a certain amount of the subject, analogously to (50), or as the
constituent structure (i.e., a certain history of object; anaiogously to (51).
derivation). However, each transformation is still
finitely characterizable (cf. § § 5.l-2), and the (50) the growling of lions
finite restricting class (41) associated with a
transformation indicates how much information
about a string is needed in order to apply this (51) the raising of flowers.
transformation. The grammar can th�refore still
be regarded as an elementary finite-state process Phrases (50) and (51), however, are not similarly
of the type corresponding to phrase structure, ambiguous. Yet in terms of phrase structure, each
There is still a bound, for each grammar, on how of these phrases is represented as: the - V '"'ing­
much of the past outpit must be inspected in order of "' NP.
for the process of derivation to continue, even
though more than just the last outpit (the last Careful analysis of English shows that we can
line of the derivation) must be known. simplify the grammar if we strike the phrases
(49)-(51) out of the kernel and reintroduce them
6. Explanatory Power of Linguistic Theories transformations.Uy by a transformation T tha.t
1
carries such sentences as "lions growl" into (50),
We have thus far considered the relative and a transformation T2 that carries such sentences
adequa.cy of theories of linguistic Btructure only

123
.. I

as "they raise flowers" into (Sl). T1 and T2 will replace "is" by "are" in (lJiii), and choose
be similar to the norninalizing transformation s5 of any required length).
described in fn.12, when they are correctly 7.z or W may be the identity element U (,cf•. fn.4)
constructed. But both "hunters shoot" and "they in this case. Note that since we limited ( 18)
shoot the hunters" are kernel sentences; and so as to exclude U from fi�ring significantl�
application of T1 to the former and T2 to the on either the right- or the left-hand side
latter yields the result (49). Hence (49) has of a rule of F, and since we reCJuired that
two distinct transformational origins. It is a only a single symbol of the left -hand side
case of constructional homonymity on the trans­ may be· replaced in any rule, it follows that
formational level. The ambiguity of the grammat­ y must be at least as long as Xi. Thus we
ical relation in (49) is a consequence of the fact 1
that the relation of "shoot" to 11 hunters 11 have a simple decision procedure for deriv­
differs in the two underlying kernel sentences. ability and terminality in the sense of
We do not have this ambiguity in the case of (SO), (l9iii), (19v).
(Sl), since neither "they growl lions" nor 8. See [J] for a detailed development of an
"flowers raise" is a grammatical kerriel sentence.
algebra of transformations for linguistic
There are many other examples of the same description and an account of transforma­
tional grammar. For further application of
general kind (cf. (l],[2]), and to my mind, they this type of description to linguistic
provide quite convincing evidence not only for
the greater adequacy of the transformational material, see (l], (2], and from a somewhat
different point of view, [4].
conception of linguistic structure, but also for
the view expressed in § 5.4 that transformational
analysis enables us to reduce partially the 9. It is not difficult to give a rigorous
problem of explaining how11e understand a sentence definition of the equivalence relation in
to that of explaining how we understand a kernel question, though this is fairly tedious.
sentence,
10. The notion 11 is a" should actually be
In summary, then, we picture a language as relativized further to a given occurrence
having a small, possibly finite kernel of basic of s in S. We can define an occurrence of
1
sentences with phrase structure in the sense of s in S as an orde:red pair (s ,X), where X
§J, along with a set of transformations which i i
can be applied to kernel sentences or to earlier is an initial substring of,S, and si is a
transforms to produce new and more complicated final substring of X. Cf. (5], p.297.
sentences from elementary components. We have ll. Where U is the identity, as in fn. 4.
seen certain indications that this approach may
enable us to reduce the immense complexity of 12. Notice that this sentence requires a
actual language to manageable proportions and-, in generalized transformation that operates on
addition, that it may provide considerable insight a pair of strings with their phrase markers.
into the actual use and understanding of language. Thus we have a transformation that converts
Footnotes s1,s2 of the forms NP-VP1, it-VP2, respect-
1. Cf. (7]. Finite-state grammars can be ively, into the string: ing ,...VP - VP 2• It
1
represented graphically by state diagrams, as converts S1= "they - prove that theorem",
in (7], p.15f. s2= 11 it - was diff icult11 in to II ing prove
2. see [6], Appendix 2, for an axiomatization of that theorem - was difficult," which by (J4)
conca.tenation algebras. becomes "proving that theorem was difficult."
Cf. [l], [J] for details.
J. By •morphemes• we refer to the smallest BibliographY:
grammatically functioning elements of the [l] Chomsky, N., The Logical Structure of
language , e. g. , 11 boy", 11 run 11 , 11 ing11 in Linguistic Theory (mimeographed).
"running", "s 11 in 11 books11, etc. [2] Chomsky, N., Syntactic Structures, to be
publ:l.shed by Mouton & Co,, 1S-Gra.venhage,
4. In the case of L1, b of (9ii) can be taken Netherlands.
j [J] Chomsky, N., Transformational Analysis,Ph. D.
as an identity element U which has the Dissertation, University of Pennsylvania,
property that for all X, U "X=X"U=X. Then June, 1955,
D will also be a dependency set for a
m [4] Hards , z.s., Discourse Analysis, Language
sentence of length 2m in L1. 28,l (1952).
[5] �uine, W.V., Mathematical Logic, revised
5. Note tha.t a. grammar must reflect and explain edition, Harvard University Press. , Cambridge,
the ability of a speaker to produce and unde:t'­ 1951.
stand new sentences which may be much longer (6) Rosenbloom, P .• Elements of Mathematical
than any he has previously heard. Logic, Dover, New York, 1950
6. Tnus we can always find sequences of n+l Shannon & Weaver, The Mathematical Theory of
words whose first n words and last n words Communication, University of Illinois Press,
may occur, but not in the same sentence (e.g. Urba.na, 1949.

124

You might also like