In linguistics, transformational grammar (TG) or transformational-generative grammar (TGG) was the earliest model of grammar proposed within the research tradition of generative grammar. [1] Like current generative theories, it treated grammar as a system of formal rules that generate all and only grammatical sentences of a given language. What was distinctive about transformational grammar was that it posited transformation rules that mapped a sentence's deep structure to its pronounced form. For example, in many variants of transformational grammar, the English active voice sentence "Emma saw Daisy" and its passive counterpart "Daisy was seen by Emma" share a common deep structure generated by phrase structure rules, differing only in that the latter's structure is modified by a passivization transformation rule.
Transformational grammar was a species of generative grammar and shared many of its goals and postulations, including the notion of linguistics as a cognitive science, the need for formal explicitness, and the competence-performance distinction. [2] Transformational grammar included two kinds of rules: phrase-structure rules and transformational rules.
Part of a series on |
Linguistics |
---|
![]() |
In transformational grammar, each sentence in a language has two levels of representation: a deep structure and a surface structure. [3] The deep structure represents a sentence's core semantic relations and is mapped onto the surface structure, which follows the sentence's phonological system very closely, via transformations.
Deep structures are generated by phrase structure grammars using rewrite rules.
Transformations are rules that map a deep structure to a surface structure. For example, a typical transformation in TG is subject-auxiliary inversion (SAI). That rule takes as its input a declarative sentence with an auxiliary, such as "John has eaten all the heirloom tomatoes", and transforms it into "Has John eaten all the heirloom tomatoes?" In the original formulation (Chomsky 1957), those rules held over strings of terminals, constituent symbols or both.
(NP = Noun Phrase and AUX = Auxiliary)
In the 1970s, by the time of the Extended Standard Theory, following Joseph Emonds's work on structure preservation, transformations came to be viewed as holding over trees. By the end of government and binding theory, in the late 1980s, transformations were no longer structure-changing operations at all; instead, they added information to already existing trees by copying constituents.
The earliest conceptions of transformations were that they were construction-specific devices. For example, there was a transformation that turned active sentences into passive ones. A different transformation raised embedded subjects into main clause subject position in sentences such as "John seems to have gone", and a third reordered arguments in the dative alternation. With the shift from rules to principles and constraints in the 1970s, those construction-specific transformations morphed into general rules (all the examples just mentioned are instances of NP movement), which eventually changed into the single general rule move alpha or Move.
Transformations actually come in two types: the post-deep structure kind mentioned above, which are string- or structure-changing, and generalized transformations (GTs). GTs were originally proposed in the earliest forms of generative grammar (such as in Chomsky 1957). They take small structures, either atomic or generated by other rules, and combine them. For example, the generalized transformation of embedding would take the kernel "Dave said X" and the kernel "Dan likes smoking" and combine them into "Dave said Dan likes smoking." GTs are thus structure-building rather than structure-changing. In the Extended Standard Theory and government and binding theory, GTs were abandoned in favor of recursive phrase structure rules, but they are still present in tree-adjoining grammar as the Substitution and Adjunction operations, and have recently reemerged in mainstream generative grammar in Minimalism, as the operations Merge and Move.
In generative phonology, another form of transformation is the phonological rule, which describes a mapping between an underlying representation (the phoneme) and the surface form that is articulated during natural speech. [4]
An important feature of all transformational grammars is that they are more powerful than context-free grammars. [5] Chomsky formalized this idea in the Chomsky hierarchy. He argued that it is impossible to describe the structure of natural languages with context-free grammars. [6] His general position on the context-dependency of natural language has held up, though his specific examples of the inadequacy of CFGs in terms of their weak generative capacity were disproved. [7] [8]
Using a term such as "transformation" may give the impression that theories of transformational generative grammar are intended as a model of the processes by which the human mind constructs and understands sentences, but Chomsky clearly stated that a generative grammar models only the knowledge that underlies the human ability to speak and understand, arguing that because most of that knowledge is innate, a baby can have a large body of knowledge about the structure of language in general and so need to learn only the idiosyncratic features of the language(s) to which it is exposed.[ citation needed ]
Chomsky is not the first person to suggest that all languages have certain fundamental things in common. He quoted philosophers who posited the same basic idea several centuries ago. But Chomsky helped make the innateness theory respectable after a period dominated by more behaviorist attitudes towards language. He made concrete and technically sophisticated proposals about the structure of language as well as important proposals about how grammatical theories' success should be evaluated. [9]
Chomsky argued that "grammatical" and "ungrammatical" can be meaningfully and usefully defined. In contrast, an extreme behaviorist linguist would argue that language can be studied only through recordings or transcriptions of actual speech and that the role of the linguist is to look for patterns in such observed speech, not to hypothesize about why such patterns might occur or to label particular utterances grammatical or ungrammatical. Few linguists in the 1950s actually took such an extreme position, but Chomsky was on the opposite extreme, defining grammaticality in an unusually mentalistic way for the time. [10] He argued that the intuition of a native speaker is enough to define the grammaticality of a sentence; that is, if a particular string of English words elicits a double-take or a feeling of wrongness in a native English speaker, with various extraneous factors affecting intuitions controlled for, it can be said that the string of words is ungrammatical. That, according to Chomsky, is entirely distinct from the question of whether a sentence is meaningful or can be understood. It is possible for a sentence to be both grammatical and meaningless, as in Chomsky's famous example, "colorless green ideas sleep furiously". [11] But such sentences manifest a linguistic problem that is distinct from that posed by meaningful but ungrammatical (non)-sentences such as "man the bit sandwich the", the meaning of which is fairly clear, but which no native speaker would accept as well-formed.
In the 1960s, Chomsky introduced two central ideas relevant to the construction and evaluation of grammatical theories.
One was the distinction between competence and performance . [12] Chomsky noted the obvious fact that when people speak in the real world, they often make linguistic errors, such as starting a sentence and then abandoning it midway through. He argued that such errors in linguistic performance are irrelevant to the study of linguistic competence, the knowledge that allows people to construct and understand grammatical sentences. Consequently, the linguist can study an idealised version of language, which greatly simplifies linguistic analysis.
The other idea related directly to evaluation of theories of grammar. Chomsky distinguished between grammars that achieve descriptive adequacy and those that go further and achieve explanatory adequacy. A descriptively adequate grammar for a particular language defines the (infinite) set of grammatical sentences in that language; that is, it describes the language in its entirety. A grammar that achieves explanatory adequacy has the additional property that it gives insight into the mind's underlying linguistic structures. In other words, it does not merely describe the grammar of a language, but makes predictions about how linguistic knowledge is mentally represented. For Chomsky, such mental representations are largely innate and so if a grammatical theory has explanatory adequacy, it must be able to explain different languages' grammatical nuances as relatively minor variations in the universal pattern of human language.
Chomsky argued that even though linguists were still a long way from constructing descriptively adequate grammars, progress in descriptive adequacy would come only if linguists held explanatory adequacy as their goal: real insight into individual languages' structure can be gained only by comparative study of a wide range of languages, on the assumption that they are all cut from the same cloth.[ citation needed ]
Chomsky developed transformational grammar in the late 1950s, drawing on older work including that of the structuralists. [13] [2] Its central ideas are maintained to varying degrees in present-day approaches to syntax such as Minimalism, while others such as Combinatory categorial grammar are distinctly non-transformational. [14]
In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). Diverse approaches, such as generative grammar and functional grammar, offer unique perspectives on syntax, reflecting its complexity and centrality to understanding human language.
In linguistics, X-bar theory is a model of phrase-structure grammar and a theory of syntactic category formation that was first proposed by Noam Chomsky in 1970 reformulating the ideas of Zellig Harris (1951), and further developed by Ray Jackendoff, along the lines of the theory of generative grammar put forth in the 1950s by Chomsky. It attempts to capture the structure of phrasal categories with a single uniform structure called the X-bar schema, basing itself on the assumption that any phrase in natural language is an XP that is headed by a given syntactic category X. It played a significant role in resolving issues that phrase structure rules had, representative of which is the proliferation of grammatical rules, which is against the thesis of generative grammar.
Deep structure and surface structure are concepts used in linguistics, specifically in the study of syntax in the Chomskyan tradition of transformational generative grammar.
Government and binding is a theory of syntax and a phrase structure grammar in the tradition of transformational grammar developed principally by Noam Chomsky in the 1980s. This theory is a radical revision of his earlier theories and was later revised in The Minimalist Program (1995) and several subsequent papers, the latest being Three Factors in Language Design (2005). Although there is a large literature on government and binding theory which is not written by Chomsky, Chomsky's papers have been foundational in setting the research agenda.
Generative grammar is a research tradition in linguistics that aims to explain the cognitive basis of language by formulating and testing explicit models of humans' subconscious grammatical knowledge. Generative linguists, or generativists, tend to share certain working assumptions such as the competence–performance distinction and the notion that some domain-specific aspects of grammar are partly innate in humans. These assumptions are rejected in non-generative approaches such as usage-based models of language. Generative linguistics includes work in core areas such as syntax, semantics, phonology, psycholinguistics, and language acquisition, with additional extensions to topics including biolinguistics and music cognition.
Principles and parameters is a framework within generative linguistics in which the syntax of a natural language is described in accordance with general principles and specific parameters that for particular languages are either turned on or off. For example, the position of heads in phrases is determined by a parameter. Whether a language is head-initial or head-final is regarded as a parameter which is either on or off for particular languages. Principles and parameters was largely formulated by the linguists Noam Chomsky and Howard Lasnik. Many linguists have worked within this framework, and for a period of time it was considered the dominant form of mainstream generative linguistics.
Syntactic Structures is an important work in linguistics by American linguist Noam Chomsky, originally published in 1957. A short monograph of about a hundred pages, it is recognized as one of the most significant and influential linguistic studies of the 20th century. It contains the now-famous sentence "Colorless green ideas sleep furiously", which Chomsky offered as an example of a grammatically correct sentence that has no discernible meaning, thus arguing for the independence of syntax from semantics.
The term linguistic performance was used by Noam Chomsky in 1960 to describe "the actual use of language in concrete situations". It is used to describe both the production, sometimes called parole, as well as the comprehension of language. Performance is defined in opposition to "competence"; the latter describes the mental knowledge that a speaker or listener has of language.
The linguistics wars were extended disputes among American theoretical linguists that occurred mostly during the 1960s and 1970s, stemming from a disagreement between Noam Chomsky and several of his associates and students. The debates started in 1967 when linguists Paul Postal, John R. Ross, George Lakoff, and James D. McCawley —self-dubbed the "Four Horsemen of the Apocalypse"—proposed an alternative approach in which the relation between semantics and syntax is viewed differently, which treated deep structures as meaning rather than syntactic objects. While Chomsky and other generative grammarians argued that meaning is driven by an underlying syntax, generative semanticists posited that syntax is shaped by an underlying meaning. This intellectual divergence led to two competing frameworks in generative semantics and interpretive semantics.
Merge is one of the basic operations in the Minimalist Program, a leading approach to generative syntax, when two syntactic objects are combined to form a new syntactic unit. Merge also has the property of recursion in that it may be applied to its own output: the objects combined by Merge are either lexical items or sets that were themselves formed by Merge. This recursive property of Merge has been claimed to be a fundamental characteristic that distinguishes language from other cognitive faculties. As Noam Chomsky (1999) puts it, Merge is "an indispensable operation of a recursive system ... which takes two syntactic objects A and B and forms the new object G={A,B}" (p. 2).
In linguistics, grammaticality is determined by the conformity to language usage as derived by the grammar of a particular speech variety. The notion of grammaticality rose alongside the theory of generative grammar, the goal of which is to formulate rules that define well-formed, grammatical sentences. These rules of grammaticality also provide explanations of ill-formed, ungrammatical sentences.
Move α is a feature of many transformational-generative grammars, first developed in the Revised Extended Standard Theory (REST) by Noam Chomsky in the late 1970s and later part of government and binding theory (GB) in the 1980s and the Minimalist Program of the 1990s. The term refers to the relation between an indexed constituent and its trace t, e.g., the relation of whom and t in the example
In linguistics, well-formedness is the quality of a clause, word, or other linguistic element that conforms to the grammar of the language of which it is a part. Well-formed words or phrases are grammatical, meaning they obey all relevant rules of grammar. In contrast, a form that violates some grammar rule is ill-formed and does not constitute part of the language.
Aspects of the Theory of Syntax is a book on linguistics written by American linguist Noam Chomsky, first published in 1965. In Aspects, Chomsky presented a deeper, more extensive reformulation of transformational generative grammar (TGG), a new kind of syntactic theory that he had introduced in the 1950s with the publication of his first book, Syntactic Structures. Aspects is widely considered to be the foundational document and a proper book-length articulation of Chomskyan theoretical framework of linguistics. It presented Chomsky's epistemological assumptions with a view to establishing linguistic theory-making as a formal discipline comparable to physical sciences, i.e. a domain of inquiry well-defined in its nature and scope. From a philosophical perspective, it directed mainstream linguistic research away from behaviorism, constructivism, empiricism and structuralism and towards mentalism, nativism, rationalism and generativism, respectively, taking as its main object of study the abstract, inner workings of the human mind related to language acquisition and production.
Lectures on Government and Binding: The Pisa Lectures (LGB) is a book by the linguist Noam Chomsky, published in 1981. It is based on the lectures Chomsky gave at the GLOW conference and workshop held at the Scuola Normale Superiore in Pisa, Italy, in 1979. In this book, Chomsky presented his government and binding theory of syntax. It had great influence on the syntactic research in early 1980s, especially among the linguists working within the transformational grammar framework.
In linguistics, Immediate Constituent Analysis (ICA) is a syntactic theory which focuses on the hierarchical structure of sentences by isolating and identifying the constituents. While the idea of breaking down sentences into smaller components can be traced back to early psychological and linguistic theories, ICA as a formal method was developed in the early 20th century. It was influenced by Wilhelm Wundt's psychological theories of sentence structure but was later refined and formalized within the framework of structural linguistics by Leonard Bloomfield. The method gained traction in the distributionalist tradition through the work of Zellig Harris and Charles F. Hockett, who expanded and applied it to sentence analysis. Additionally, ICA was further explored within the context of glossematics by Knud Togeby. These contributions helped ICA become a central tool in syntactic analysis, focusing on the hierarchical relationships between sentence constituents.
In linguistics, the term formalism is used in a variety of meanings which relate to formal linguistics in different ways. In common usage, it is merely synonymous with a grammatical model or a syntactic model: a method for analyzing sentence structures. Such formalisms include different methodologies of generative grammar which are especially designed to produce grammatically correct strings of words; or the likes of Functional Discourse Grammar which builds on predicate logic.
In linguistics, transformational syntax is a derivational approach to syntax that developed from the extended standard theory of generative grammar originally proposed by Noam Chomsky in his books Syntactic Structures and Aspects of the Theory of Syntax. It emerged from a need to improve on approaches to grammar in structural linguistics.
Current Issues in Linguistic Theory is a 1964 book by American linguist Noam Chomsky. It is a revised and expanded version of "The Logical Basis of Linguistic Theory", a paper that Chomsky presented in the ninth International Congress of Linguists held in Cambridge, Massachusetts in 1962. It is a short monograph of about a hundred pages, similar to Chomsky's earlier Syntactic Structures (1957). In Aspects of the Theory of Syntax (1965), Chomsky presents many of its ideas in a more elaborate manner.
The lexicalist hypothesis is a hypothesis proposed by Noam Chomsky in which he claims that syntactic transformations only can operate on syntactic constituents. It says that the system of grammar that assembles words is separate and different from the system of grammar that assembles phrases out of words.
Early generative work was known as "transformational grammar"