Aspects of the Theory of Syntax

Last updated
Aspects of the Theory of Syntax
AspectsOfTheTheoryOfSyntax.jpg
Cover of the first edition
Author Noam Chomsky
LanguageEnglish
Subject Natural language syntax
Publisher MIT Press
Publication date
May 1965
Media typePrint
Pages261
Preceded by Syntactic Structures  
Followed by Cartesian linguistics  

Aspects of the Theory of Syntax (known in linguistic circles simply as Aspects [1] ) is a book on linguistics written by American linguist Noam Chomsky, first published in 1965. In Aspects, Chomsky presented a deeper, more extensive reformulation of transformational generative grammar (TGG), a new kind of syntactic theory that he had introduced in the 1950s with the publication of his first book, Syntactic Structures . Aspects is widely considered to be the foundational document and a proper book-length articulation of Chomskyan theoretical framework of linguistics. [2] It presented Chomsky's epistemological assumptions with a view to establishing linguistic theory-making as a formal (i.e. based on the manipulation of symbols and rules) discipline comparable to physical sciences, i.e. a domain of inquiry well-defined in its nature and scope. From a philosophical perspective, it directed mainstream linguistic research away from behaviorism, constructivism, empiricism and structuralism and towards mentalism, nativism, rationalism and generativism, respectively, taking as its main object of study the abstract, inner workings of the human mind related to language acquisition and production.

Contents

Background

After the publication of Chomsky's Syntactic Structures , the nature of linguistic research began to change, especially at MIT and elsewhere in the linguistic community where TGG had a favorable reception. Morris Halle, a student of Roman Jacobson and a colleague of Chomsky at MIT's Research Laboratory of Electronics (RLE), was a strong supporter of Chomsky's ideas of TGG. At first Halle worked on a generative phonology of Russian and published his work in 1959. [3] From 1956 until 1968, together with Chomsky (and also with Fred Lukoff initially), Halle developed a new theory of phonology called generative phonology. Their collaboration culminated with the publication of The Sound Pattern of English in 1968. Robert Lees, a linguist of the traditional structuralist school, went to MIT in 1956 to work in the mechanical translation project at RLE, but became convinced by Chomsky's TGG approach and went on to publish, in 1960, probably the very first book of a linguistic analysis based on TGG entitled The Grammar of English Nominalizations. This work was preceded by Lees's doctoral thesis on the same topic, for which he was given a Ph.D. in electrical engineering. Lees was technically the first student of the new TGG paradigm. Edward S. Klima, a graduate of the Masters program from Harvard and hired by Chomsky at RLE in 1957, produced pioneering TGG-based work on negation. [4] In 1959, Chomsky wrote a critical review of B. F. Skinner's Verbal Behavior (1957) in the journal Language , in which he emphasized on the fundamentally human characteristic of verbal creativity, which is present even in very young children, and rejected the behaviorist way of describing language in ambiguous terms such as "stimulus", "response", "habit", "conditioning", and "reinforcement". [5]

With Morris Halle and others, Chomsky founded the graduate program in linguistics at MIT in 1961. The program immediately attracted some of the brightest young American linguists. Jerry Fodor and Jerrold Katz, both graduates of the Ph.D. program at Princeton, and Paul Postal, a Ph.D. from Yale, were some of the first students of this program. They made major contributions to the nascent field of TGG. John Viertel, a colleague of Chomsky at RLE in the 1950s, began working for a Ph.D. dissertation under Chomsky on the linguistic thoughts of Wilhelm von Humboldt, a nineteenth-century German linguist. Viertel's English translations of Humboldt's works influenced Chomsky at this time and made him abandon Saussurian views of linguistics. [6] Chomsky also collaborated with visiting French mathematician Marcel-Paul Schützenberger, and was able to formulate one of the most important theorems of formal linguistics, the Chomsky-Schützenberger hierarchy. Within the theoretical framework of TGG, G. H. Matthews, Chomsky's colleague at RLE, worked on the grammar of Hidatsa, a Native American language. J. R. Applegate worked on the German noun phrase. Lees and Klima looked into English pronominalization. Matthews and Lees worked on the German verb phrase. [7] On the nature of the linguistic research at MIT in those days, Jerry Fodor recalls that "...communication was very lively, and I guess we shared a general picture of the methodology for doing, not just linguistics, but behavioral science research. We were all more or less nativist, and all more or less mentalist. There was a lot of methodological conversation that one didn't need to have. One could get right to the substantive issues. So, from that point of view, it was extremely exciting." [8] In 1962, Chomsky gave a paper at the Ninth International Congress of Linguists entitled "The Logical Basis of Linguistic Theory", in which he outlined the transformational generative grammar approach to linguistics. In June 1964, he delivered a series of lectures at the Linguistic Institute of the Linguistic Society of America (these were later published in 1966 as Topics in the Theory of Generative Grammar).

All of these activities aided to develop what is now known as the "Standard Theory" of TGG, in which the basic formulations of Syntactic Structures underwent considerable revision. In 1965, eight years after the publication of Syntactic Structures, Chomsky published Aspects partly as an acknowledgment of this development and partly as a guide for future directions for the field.

Overview of topics

As British linguist Peter Hugoe Matthews noted in his review [9] of the book, the content of Aspects can be divided into two distinct parts: Chapter 1 is concerned with the psychological reality of language and the philosophy of language research, and the rest of the chapters deal with specific technical details within generative grammar.

The goal of linguistic theory

Competence vs. performance: descriptive adequacy

In Aspects, Chomsky lays down the abstract, idealized context in which a linguistic theorist is supposed to perform his research: "Linguistic theory is concerned primarily with an ideal speaker-listener, in a completely homogeneous speech-community, who knows its language perfectly and is unaffected by such grammatically irrelevant conditions as memory limitations, distractions, shifts of attention and interest, and errors (random or characteristic) in applying his knowledge of the language in actual performance." He makes a "fundamental distinction between competence (the speaker-hearer's knowledge of his language) and performance (the actual use of language in concrete situation)." [10] A "grammar of a language" is "a description of the ideal speaker-hearer's intrinsic competence", and this "underlying competence" is a "system of generative processes." [11] An "adequate grammar" should capture the basic regularities and the productive nature of a language. [12] Chomsky calls this "descriptive adequacy" of the linguistic theory, in the sense that "it correctly describes its object, namely the linguistic intuition—the tacit competence—of the native speaker. In this sense, the grammar is justified on external grounds, on grounds of correspondence to linguistic fact." [13]

Language acquisition, universal grammar and explanatory adequacy

Additionally, Chomsky sets forth another ambitious goal for linguistic theory in Aspects: that it has to be "sufficiently rich to account for acquisition of language, yet not so rich as to be inconsistent with the known diversity of language." In other words, linguistic theory must be able to describe how any normal human child masters the complexities of his mother tongue at such a young age, and how children all over the world master languages which are enormously different from one another in terms of vocabulary, word order and morpho-syntactic constructions.

In Chomsky's opinion, in order for a linguistic theory to be justified on "internal grounds" and to achieve "explanatory adequacy", it has to show how a child's brain, when exposed to primary linguistic data, uses special innate abilities or strategies (described as a set of principles called "Universal Grammar") and selects the correct grammar of the language over many other grammars compatible with the same data. [13]

Grammaticality and acceptability

For Chomsky, "grammaticalness is ... a matter of degree." [14] When sentences are directly generated by the system of grammatical rules, they are called "perfectly" or "strictly well-formed" grammatical sentences. [15] When sentences are "derivatively generated" by "relaxing" some grammatical rules (such as "subcategorization rules" or "selectional rules" [15] ), they deviate from strictly well-formedness. [16] Chomsky calls these grammatically "deviant". [17] The degree and manner of their deviation can be evaluated by comparing their structural description with that of the strictly well-formed sentences. In this way, a theory of "degree of grammaticalness" can eventually be developed. [15]

According to Chomsky, an "acceptable" sentence is one that is "perfectly natural" and "immediately comprehensible" and "in no way bizarre or outlandish". [18] The notion of acceptability depends on various "dimensions" such as "rapidity, correctness, and uniformity of recall and recognition, normalcy of intonation". [18] Chomsky adds that "acceptability is a concept that belongs to the study of performance, whereas grammaticalness belongs to the study of competence." [14] So, there can be sentences that are grammatical but nevertheless unacceptable because of "memory limitations" or intonational and stylistic factors." [14]

Emphasis on mentalism

In Aspects Chomsky writes that "linguistic theory is mentalistic, since it is concerned with discovering a mental reality underlying actual behavior." [11] With this mentalist interpretation of linguistic theory, Chomsky elevated linguistics to a field that is part of a broader theory of human mind, i.e. the cognitive sciences. According to Chomsky, a human child's mind is equipped with a "language acquisition device" formed by inborn mental properties called "linguistic universals" which eventually constructs a mental theory of the child's mother tongue. [19] The linguist's main object of inquiry, as Chomsky sees it, is this underlying psychological reality of language. Instead of making catalogs and summaries of linguistic behavioral data demonstrated on the surface (i.e. behaviorism), a Chomskyan linguist should be interested in using "introspective data" to ascertain the properties of a deeper mental system.

The mentalist approach to linguistics proposed by Chomsky is also different from an investigation of the neurophysiological mechanisms underlying language. It is about abstractly determining the properties and functions of such mechanisms. [20]

The structure of grammar: deep structure

The grammar model discussed in Noam Chomsky's Aspects of the Theory of Syntax (1965) Aspects Grammar Model.jpg
The grammar model discussed in Noam Chomsky's Aspects of the Theory of Syntax (1965)

In Aspects, Chomsky summarized his proposed structure of a grammar in the following way: "A grammar contains a syntactic component, a semantic component and a phonological component...The syntactic component consists of a base and a transformational component. The base, in turn, consists of a categorial subcomponent and a lexicon. The base generates deep structures. A deep structure enters the semantic component and receives a semantic interpretation; it is mapped by transformational rules into a surface structure, which is then given a phonetic interpretation by the rules of the phonological component." [21] In this grammar model, syntax is given a prominent, generative role, whereas phonology and semantics are assigned secondary, interpretive roles. This theory of grammar would later come to be known as the "Standard Theory" (ST).

The base subcomponent

The base in the syntactic component functions as follows: In the first step, a simple set of phrase structure rules generate tree diagrams (sometimes called Phrase Markers) consisting of nodes and branches, but with empty terminal nodes; these are called "pre-lexical structures". In the second step, the empty terminal nodes are filled with complex symbols consisting of morphemes accompanied by syntactic and semantic features, supplied from the lexicon via lexical insertion rules. The resulting tree diagram is called a "deep structure".

Comparison with the Syntactic Structures model

The Aspects model or ST differed from Syntactic Structures (1957) in a number of ways. Firstly, the notion of kernel sentences (a class of sentences produced by applying obligatory transformational rules) was abandoned and replaced by the notion of "deep structures", within which negative, interrogative markers, etc. are embedded. This simplified the generation of "surface" sentences, whereas in the previous model, a number of successive optional transformational rules had to be applied on the kernel sentences to arrive at the same result.

Secondly, the addition of a semantic component to the grammar marked an important conceptual change since Syntactic Structures, where the role of meaning was effectively neglected and not considered part of the grammatical model. [22] Chomsky mentions that the semantic component is essentially the same as described in Katz and Postal (1964).

Among the more technical innovations are the use of recursive phrase structure rules and the introduction of syntactic features in lexical entries to address the issue of subcategorization.

Syntactic features

In Chapter 2 of Aspects, Chomsky discusses the problem of subcategorization of lexical categories and how this information should be captured in a generalized manner in the grammar. He deems that rewriting rules are not the appropriate device in this regard. As a solution, he borrows the idea of features use in phonology. A lexical category such as noun, verb, etc. is represented by a symbol such as N, V. etc. A set of "subcategorization rules" then analyzes these symbols into "complex symbols", each complex symbol being a set of specified "syntactic features", grammatical properties with binary values.

Syntactic feature is one of the most important technical innovations of the Aspects model. Most contemporary grammatical theories have preserved it.

Significance

Linguistics

UCLA linguist Tim Stowell considers Aspects to be "effectively the most important foundational document of the field" of transformational generative grammar (TGG), providing "the definitive exposition of the classical theory of TGG—the so-called Standard Theory". [2]

University of Cambridge linguists Ian Roberts and Jeffrey Watumull maintain that Aspects ushered in the "Second Cognitive Revolution—the revival of rationalist philosophy first expounded in the Enlightenment", in particular by Leibniz. [23]

Philosophy

Moral philosopher John Rawls compared building a theory of morality to that of the generative grammar model found in Aspects. In A Theory of Justice (1971), he notes that just like Chomsky's grammar model assumes a set of finite underlying principles that are supposed to adequately explain the variety of sentences in linguistic performance, our sense of justice can be defined as a set of moral principles that give rise to everyday judgments. [24]

Medicine

In his Nobel Prize lecture titled "The Generative Grammar of the Immune System", the 1984 Nobel Prize laureate in Medicine and Physiology Niels K. Jerne used Chomsky's generative grammar model in Aspects to explain the human immune system, comparing "the variable region of a given antibody molecule" to "a sentence". "The immense repertoire of the immune system then becomes ... a lexicon of sentences which is capable of responding to any sentence expressed by the multitude of antigens which the immune system may encounter." Jerne called the DNA segments in chromosomes which encode the variable regions of antibody polypeptides a human's inheritable "deep structures", which can account for the innately complex yet miraculously effective fighting capacity of human antibodies against complex antigens. This is comparable to Chomsky's hypothesis that a child is born with innate language universals to acquire a complex human language in a very short time. [25]

Artificial intelligence

Neuroscientist David Marr wrote that the goal of artificial intelligence is to find and solve information processing problems. First, one must build a computational theory of the problem (i.e. the abstract formulation of the "what" and "why" of the problem). And then one must construct an algorithm that implements it (i.e. the "how" of the problem). Marr likened the computational theory of an information processing problem to the notion of "competence" mentioned in Aspects. [26]

Criticism

Several of the theoretical constructs and principles of the generative grammar introduced in Aspects such as deep structures, transformations, autonomy and primacy of syntax, etc. were either abandoned or substantially revised after they were shown to be either inadequate or too complicated to account for, in a simple and elegant way, many idiosyncratic example sentences from different languages. [27] As a response to these problems encountered within the Standard Theory, a new approach called the generative semantics (as opposed to the interpretive semantics in Aspects) was invented in the early 1970s by some of Chomsky's collaborators (notably George Lakoff [28] ), and was incorporated later in the late 1980s into what is now known as the school of cognitive linguistics, at odds with Chomskyan school of generative linguistics. Chomsky himself addressed these issues at around the same time (early 1970s) and updated the model to an "Extended Standard Theory", where syntax was less autonomous, the interaction between the syntactic and the semantic component was much more interactive and the transformations were cyclical. [29]

Related Research Articles

The following outline is provided as an overview and topical guide to linguistics:

<span class="mw-page-title-main">Syntax</span> System responsible for combining morphemes into complex structures

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). There are numerous approaches to syntax that differ in their central assumptions and goals.

In linguistics, transformational grammar (TG) or transformational-generative grammar (TGG) is part of the theory of generative grammar, especially of natural languages. It considers grammar to be a system of rules that generate exactly those combinations of words that form grammatical sentences in a given language and involves the use of defined operations to produce new sentences from existing ones.

Lexical semantics, as a subfield of linguistic semantics, is the study of word meanings. It includes the study of how words structure their meaning, how they act in grammar and compositionality, and the relationships between the distinct senses and uses of a word.

In linguistics, X-bar theory is a model of phrase-structure grammar and a theory of syntactic category formation that was first proposed by Noam Chomsky in 1970 reformulating the ideas of Zellig Harris (1951), and further developed by Ray Jackendoff, along the lines of the theory of generative grammar put forth in the 1950s by Chomsky. It attempts to capture the structure of phrasal categories with a single uniform structure called the X-bar schema, basing itself on the assumption that any phrase in natural language is an XP that is headed by a given syntactic category X. It played a significant role in resolving issues that phrase structure rules had, representative of which is the proliferation of grammatical rules, which is against the thesis of generative grammar.

Deep structure and surface structure are concepts used in linguistics, specifically in the study of syntax in the Chomskyan tradition of transformational generative grammar.

<span class="mw-page-title-main">Generative grammar</span> Theory in linguistics

Generative grammar, or generativism, is a linguistic theory that regards linguistics as the study of a hypothesised innate grammatical structure. It is a biological or biologistic modification of earlier structuralist theories of linguistics, deriving from logical syntax and glossematics. Generative grammar considers grammar as a system of rules that generates exactly those combinations of words that form grammatical sentences in a given language. It is a system of explicit rules that may apply repeatedly to generate an indefinite number of sentences which can be as long as one wants them to be. The difference from structural and functional models is that the object is base-generated within the verb phrase in generative grammar. This purportedly cognitive structure is thought of as being a part of a universal grammar, a syntactic structure which is caused by a genetic mutation in humans.

<i>Syntactic Structures</i> Book by Noam Chomsky

Syntactic Structures is an important work in linguistics by American linguist Noam Chomsky, originally published in 1957. A short monograph of about a hundred pages, it is recognized as one of the most significant and influential linguistic studies of the 20th century. It contains the now-famous sentence "Colorless green ideas sleep furiously", which Chomsky offered as an example of a grammatically correct sentence that has no discernible meaning, thus arguing for the independence of syntax from semantics.

In linguistics, linguistic competence is the system of unconscious knowledge that one knows when they know a language. It is distinguished from linguistic performance, which includes all other factors that allow one to use one's language in practice.

Generative semantics was a research program in theoretical linguistics which held that syntactic structures are computed on the basis of meanings rather than the other way around. Generative semantics developed out of transformational generative grammar in the mid-1960s, but stood in opposition to it. The period in which the two research programs coexisted was marked by intense and often personal clashes now known as the linguistics wars. Its proponents included Haj Ross, Paul Postal, James McCawley, and George Lakoff, who dubbed themselves "The Four Horsemen of the Apocalypse".

In linguistics and social sciences, markedness is the state of standing out as nontypical or divergent as opposed to regular or common. In a marked–unmarked relation, one term of an opposition is the broader, dominant one. The dominant default or minimum-effort form is known as unmarked; the other, secondary one is marked. In other words, markedness involves the characterization of a "normal" linguistic unit against one or more of its possible "irregular" forms.

The term linguistic performance was used by Noam Chomsky in 1960 to describe "the actual use of language in concrete situations". It is used to describe both the production, sometimes called parole, as well as the comprehension of language. Performance is defined in opposition to "competence"; the latter describes the mental knowledge that a speaker or listener has of language.

The linguistics wars were a protracted academic dispute inside American theoretical linguistics that took place mostly in the 1960s and 1970s, stemming from an intellectual falling out between Noam Chomsky and some of his early colleagues and doctoral students. The debate began in 1967, when linguists Paul Postal, "Haj" Ross, George Lakoff, and James McCawley—self-dubbed the "Four Horsemen of the Apocalypse" —proposed an approach to the relationship between syntax and semantics, which treated deep structures as meanings rather than syntactic objects. While Chomsky and other generative grammarians argued that the meaning of a sentence was derived from its syntax, the generative semanticists argued that syntax was derived from meaning.

In linguistics, well-formedness is the quality of a clause, word, or other linguistic element that conforms to the grammar of the language of which it is a part. Well-formed words or phrases are grammatical, meaning they obey all relevant rules of grammar. In contrast, a form that violates some grammar rule is ill-formed and does not constitute part of the language.

<i>Lectures on Government and Binding</i> 1981 book by Noam Chomsky

Lectures on Government and Binding: The Pisa Lectures (LGB) is a book by the linguist Noam Chomsky, published in 1981. It is based on the lectures Chomsky gave at the GLOW conference and workshop held at the Scuola Normale Superiore in Pisa, Italy, in 1979. In this book, Chomsky presented his government and binding theory of syntax. It had great influence on the syntactic research in early 1980s, especially among the linguists working within the transformational grammar framework.

<span class="mw-page-title-main">Formalism (linguistics)</span> Concept in linguistics

In linguistics, the term formalism is used in a variety of meanings which relate to formal linguistics in different ways. In common usage, it is merely synonymous with a grammatical model or a syntactic model: a method for analyzing sentence structures. Such formalisms include different methodologies of generative grammar which are especially designed to produce grammatically correct strings of words; or the likes of Functional Discourse Grammar which builds on predicate logic.

In linguistics, transformational syntax is a derivational approach to syntax that developed from the extended standard theory of generative grammar originally proposed by Noam Chomsky in his books Syntactic Structures and Aspects of the Theory of Syntax. It emerged from a need to improve on approaches to grammar in structural linguistics.

<i>Current Issues in Linguistic Theory</i> 1964 book by Noam Chomsky

Current Issues in Linguistic Theory is a 1964 book by American linguist Noam Chomsky. It is a revised and expanded version of "The Logical Basis of Linguistic Theory", a paper that Chomsky presented in the ninth International Congress of Linguists held in Cambridge, Massachusetts in 1962. It is a short monograph of about a hundred pages, similar to Chomsky's earlier Syntactic Structures (1957). In Aspects of the Theory of Syntax (1965), Chomsky presents many of its ideas in a more elaborate manner.

Lexicalist hypothesis is a hypothesis proposed by Noam Chomsky in which he claims that syntactic transformations only can operate on syntactic constituents. The hypothesis states that the system of grammar that assembles words is separate and different from the system of grammar that assembles phrases out of words.

In linguistics, the autonomy of syntax is the assumption that syntax is arbitrary and self-contained with respect to meaning, semantics, pragmatics, discourse function, and other factors external to language. The autonomy of syntax is advocated by linguistic formalists, and in particular by generative linguistics, whose approaches have hence been called autonomist linguistics.

References

  1. See Gallego and Ott 2015
  2. 1 2 Gallego and Ott 2015 : 249
  3. Halle 1959
  4. Klima 1964
  5. Chomsky 1959
  6. In Chomsky 1970:23, Chomsky writes that "Modern linguistics is much under the influence of Saussure's conception of langue as an inventory of elements and his preoccupation with systems of elements rather than the systems of rules which were the focus of attention in traditional grammar and in the general linguistics of Humboldt."
  7. Yngve 1956.
  8. Harris 1993: 68
  9. Matthews 1967
  10. Chomsky 1965 , p. 3
  11. 1 2 Chomsky 1965 , p. 4
  12. Chomsky 1965 , pp. 4–5
  13. 1 2 Chomsky 1965 , pp. 26–27
  14. 1 2 3 Chomsky 1965 , p. 11
  15. 1 2 3 Chomsky 1965 , p. 150
  16. Chomsky 1965 , p. 227
  17. Chomsky 1965 , p. 149
  18. 1 2 Chomsky 1965 , p. 10
  19. Chomsky 1965: 33
  20. Chomsky 1965: 193
  21. Chomsky 1965: 141
  22. From Chomsky 1957:106 : "Grammar is best formulated as a self-contained study independent of semantics."
  23. Gallego and Ott 2015 : 211
  24. Rawls 1971:41-42
  25. Jerne 1985
  26. Marr 1977
  27. For example, in Smith and Wilson 1979: 122, it was argued that “deep structure can be abandoned, not because it is not deep enough to handle all the facts, but because it is deeper than is necessary.”
  28. In Lakoff 1973, Lakoff writes that "[Generative Semanticists] have found that many of the most basic assumptions of transformational grammar were inadequate and have rejected them, including the following of Chomsky’s fundamental assumptions: that syntax is independent of human thought and reasoning, that there exists a syntactic deep structure, that transformational rules are fundamentally adequate for the study of grammar, that syntactic categories are independent of the categories of human thought, that language use plays no role in grammar, that syntax is independent of the social and cultural assumptions of speakers, and many other central positions of Chomsky’s that many of us find inadequate, especially in the light of recent research."
  29. Chomsky 1973

Bibliography