Grammatical features |
---|
In linguistics, branching refers to the shape of the parse trees that represent the structure of sentences. [1] Assuming that the language is being written or transcribed from left to right, parse trees that grow down and to the right are right-branching, and parse trees that grow down and to the left are left-branching. The direction of branching reflects the position of heads in phrases, and in this regard, right-branching structures are head-initial , whereas left-branching structures are head-final. [2] English has both right-branching (head-initial) and left-branching (head-final) structures, although it is more right-branching than left-branching. [3] Some languages such as Japanese and Turkish are almost fully left-branching (head-final). [4] Some languages are mostly right-branching (head-initial).
Languages typically construct phrases with a head word (or nucleus) and zero or more dependents (modifiers). The following phrases show the phrase heads in bold.
Examples of left-branching phrases (= head-final phrases):
Examples of right-branching phrases (= head-initial phrases):
Examples of phrases that contain both left- and right-branching (= head-medial phrases):
Concerning phrases such as the house and the house there, this article assumes the traditional NP analysis, meaning that the noun is deemed to be head over the determiner. On a DP-analysis (determiner phrase), the phrase the house would be right-branching instead of left-branching.
Left- and right-branching structures are illustrated with the trees that follow. Each example appears twice, once according to a constituency-based analysis associated with a phrase structure grammar [5] and once according to a dependency-based analysis associated with a dependency grammar. [6] The first group of trees illustrate left-branching:
The upper row shows the constituency-based structures, and the lower row the dependency-based structures. In the constituency-based structures, left-branching is present (but not really visible) in so far as the non-head daughter is to the left of the head. In the corresponding dependency-based structures in the lower row, the left-branching is clear; the dependent appears to the left of its head, the branch extending down to the left. The following structures demonstrate right-branching:
The upper row again shows the constituency-based structures, and the lower row the dependency-based structures. The constituency-based structures are right-branching insofar as the non-head daughter is to the right of the head. This right-branching is completely visible in the lower row of dependency-based structures, where the branch extends down to the right. The (c)-examples contain one instance of right-branching (the upper branch) and one instance of left-branching (the lower branch). The following trees illustrate phrases that combine both types of branching:
The combination of left- and right-branching is now completely visible in both the constituency- and dependency-based trees. The head appears in a medial position, which means that the phrase combines both types of branching. Note that the (b)-trees also contain a prepositional phrase that is an instance of pure right-branching.
The nature of branching is most visible with full trees. The following trees have been chosen to illustrate the extent to which a structure can be entirely left- or entirely right-branching. The following sentence is completely left-branching. The constituency-based trees are on the left, and the dependency-based trees are on the right: [7]
The category Po (= possessive) is used to label possessive 's. The following sentence is completely right-branching:
Most structures in English are, however, not completely left- or completely right-branching, but rather they combine both. The following trees illustrate what can be seen as a stereotypical combination of left- and right-branching in English:
Determiners (e.g. the) always and subjects (e.g. the child) usually appear on left branches in English, but infinitival verbs (e.g. try, eat) and the verb particle to usually appear on right branches. In the big picture, right-branching structures tend to outnumber the left-branching structures in English, which means that trees usually grow down to the right.
The X-bar schema [8] combines left- and right-branching. The standard X-bar schema has the following structure:
This structure is both left- and right branching. It is left-branching insofar as the bar-level projection of the head (X') follows the specifier, but it is right-branching insofar as the actual head (X0) precedes the complement. Despite these conflicting traits, most standard X-bar structures (in English) are more right-branching than left-branching because specifiers tend to be less complex (i.e. fewer words) than complements.
Much work in Government and Binding Theory (GB), the Minimalist Program (MP), and Lexical Functional Grammar (LFG) assumes all branching to be binary. [9] Other theories (both constituency- and dependency-based ones), e.g. early Transformational Grammar, Head-Driven Phrase Structure Grammar, Meaning-Text Theory, Word Grammar, etc. allow for n-ary branching. This distinction can have a profound impact on the overall nature of the theory of syntax. The two main possibilities in a phrase structure grammar are illustrated with the following trees:
The binary branching on the left is closely associated with the structures of GB, MP, and LFG, and it is similar to what the X-bar schema assumes. The n-ary branching structure on the right is a more traditional approach to branching. One can muster arguments for both approaches. For instance, the critics of the strictly binary branching structures charge that the strict binarity is motivated more by a desire for theoretical purity than by empirical observation.
Strictly binary branching structures increase the amount of syntactic structure (number of nodes) to the upper limit of what is possible, whereas flatter n-ary branching tends to restrict the amount of structure that the theory can assume. Worth noting in this area is that the more layered the syntactic structures are, the more discontinuities can occur, which means the component of the theory that addresses discontinuities must play a greater role. Given the flatter structures associated with n-ary branching, certain phenomena (e.g. inversion and shifting) do not result in discontinuities, a fact that reduces the role that the component for discontinuities must play in the theory.
As stated above, the main branching trait for a language is just a tendency and it often shows exceptions. Spanish, for example, while overwhelmingly right-branching, puts numeral modifiers before nouns and, in certain cases, objects before verbs. Languages like English or Swedish, though regarded as being right-branching because the main verbs precede direct objects, place adjectives and numerals before their nouns. On the contrary, Northwest Caucasian languages such as Ubykh or Abkhaz are strongly left-branching but put adjectives after nouns.
Some languages, however, are exclusively left-branching or right-branching. Japanese and most other languages of northeastern Asia and the Indian subcontinent, as well as the Turkic languages, are practically a model for rigidly left-branching languages. The Mon–Khmer and Austronesian languages of southeast Asia and many African languages come close to rigidly right-branching, with numerals as well as adjectives following their nouns and with degree words like very, too, extremely, and quite following the adjectives they modify.
In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). Diverse approaches, such as generative grammar and functional grammar, offer unique perspectives on syntax, reflecting its complexity and centrality to understanding human language.
A syntactic category is a syntactic unit that theories of syntax assume. Word classes, largely corresponding to traditional parts of speech, are syntactic categories. In phrase structure grammars, the phrasal categories are also syntactic categories. Dependency grammars, however, do not acknowledge phrasal categories.
In grammar, a phrase—called expression in some contexts—is a group of words or singular word acting as a grammatical unit. For instance, the English expression "the very happy squirrel" is a noun phrase which contains the adjective phrase "very happy". Phrases can consist of a single word or a complete sentence. In theoretical linguistics, phrases are often analyzed as units of syntactic structure such as a constituent. There is a difference between the common use of the term phrase and its technical use in linguistics. In common usage, a phrase is usually a group of words with some special idiomatic meaning or other significance, such as "all rights reserved", "economical with the truth", "kick the bucket", and the like. It may be a euphemism, a saying or proverb, a fixed expression, a figure of speech, etc.. In linguistics, these are known as phrasemes.
Phrase structure rules are a type of rewrite rule used to describe a given language's syntax and are closely associated with the early stages of transformational grammar, proposed by Noam Chomsky in 1957. They are used to break down a natural language sentence into its constituent parts, also known as syntactic categories, including both lexical categories and phrasal categories. A grammar that uses phrase structure rules is a type of phrase structure grammar. Phrase structure rules as they are commonly employed operate according to the constituency relation, and a grammar that employs phrase structure rules is therefore a constituency grammar; as such, it stands in contrast to dependency grammars, which are based on the dependency relation.
A noun phrase – or NP or nominal (phrase) – is a phrase that usually has a noun or pronoun as its head, and has the same grammatical functions as a noun. Noun phrases are very common cross-linguistically, and they may be the most frequently occurring phrase type.
A parse tree or parsing tree is an ordered, rooted tree that represents the syntactic structure of a string according to some context-free grammar. The term parse tree itself is used primarily in computational linguistics; in theoretical syntax, the term syntax tree is more common.
In linguistics, X-bar theory is a model of phrase-structure grammar and a theory of syntactic category formation that was first proposed by Noam Chomsky in 1970 reformulating the ideas of Zellig Harris (1951), and further developed by Ray Jackendoff, along the lines of the theory of generative grammar put forth in the 1950s by Chomsky. It attempts to capture the structure of phrasal categories with a single uniform structure called the X-bar schema, basing itself on the assumption that any phrase in natural language is an XP that is headed by a given syntactic category X. It played a significant role in resolving issues that phrase structure rules had, representative of which is the proliferation of grammatical rules, which is against the thesis of generative grammar.
In linguistics, a determiner phrase (DP) is a type of phrase headed by a determiner such as many. Controversially, many approaches take a phrase like not very many apples to be a DP, headed, in this case, by the determiner many. This is called the DP analysis or the DP hypothesis. Others reject this analysis in favor of the more traditional NP analysis where apples would be the head of the phrase in which the DP not very many is merely a dependent. Thus, there are competing analyses concerning heads and dependents in nominal groups. The DP analysis developed in the late 1970s and early 1980s, and it is the majority view in generative grammar today.
In linguistics, the head or nucleus of a phrase is the word that determines the syntactic category of that phrase. For example, the head of the noun phrase boiling hot water is the noun water. Analogously, the head of a compound is the stem that determines the semantic category of that compound. For example, the head of the compound noun handbag is bag, since a handbag is a bag, not a hand. The other elements of the phrase or compound modify the head, and are therefore the head's dependents. Headed phrases and compounds are called endocentric, whereas exocentric ("headless") phrases and compounds lack a clear head. Heads are crucial to establishing the direction of branching. Head-initial phrases are right-branching, head-final phrases are left-branching, and head-medial phrases combine left- and right-branching.
Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.
The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue. Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining character of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.
A sentence diagram is a pictorial representation of the grammatical structure of a sentence. The term "sentence diagram" is used more when teaching written language, where sentences are diagrammed. The model shows the relations between words and the nature of sentence structure and can be used as a tool to help recognize which potential sentences are actual sentences.
In theoretical linguistics, a distinction is made between endocentric and exocentric constructions. A grammatical construction is said to be endocentric if it fulfils the same linguistic function as one of its parts, and exocentric if it does not. The distinction reaches back at least to Bloomfield's work of the 1930s, who based it on terms by Pāṇini and Patañjali in Sanskrit grammar. Such a distinction is possible only in phrase structure grammars, since in dependency grammars all constructions are necessarily endocentric.
Lucien Tesnière was a prominent and influential French linguist. He was born in Mont-Saint-Aignan on May 13, 1893. As a senior lecturer at the University of Strasbourg (1924) and later professor at the University of Montpellier (1937), he published many papers and books on Slavic languages. However, his importance in the history of linguistics is based mainly on his development of an approach to the syntax of natural languages that would become known as dependency grammar. He presented his theory in his book Éléments de syntaxe structurale, published posthumously in 1959. In the book he proposes a sophisticated formalization of syntactic structures, supported by many examples from a diversity of languages. Tesnière died in Montpellier on December 6, 1954.
In linguistics, an argument is an expression that helps complete the meaning of a predicate, the latter referring in this context to a main verb and its auxiliaries. In this regard, the complement is a closely related concept. Most predicates take one, two, or three arguments. A predicate and its arguments form a predicate-argument structure. The discussion of predicates and arguments is associated most with (content) verbs and noun phrases (NPs), although other syntactic categories can also be construed as predicates and as arguments. Arguments must be distinguished from adjuncts. While a predicate needs its arguments to complete its meaning, the adjuncts that appear with a predicate are optional; they are not necessary to complete the meaning of the predicate. Most theories of syntax and semantics acknowledge arguments and adjuncts, although the terminology varies, and the distinction is generally believed to exist in all languages. Dependency grammars sometimes call arguments actants, following Lucien Tesnière (1959).
Merge is one of the basic operations in the Minimalist Program, a leading approach to generative syntax, when two syntactic objects are combined to form a new syntactic unit. Merge also has the property of recursion in that it may be applied to its own output: the objects combined by Merge are either lexical items or sets that were themselves formed by Merge. This recursive property of Merge has been claimed to be a fundamental characteristic that distinguishes language from other cognitive faculties. As Noam Chomsky (1999) puts it, Merge is "an indispensable operation of a recursive system ... which takes two syntactic objects A and B and forms the new object G={A,B}" (p. 2).
In linguistics, Immediate Constituent Analysis (ICA) is a syntactic theory which focuses on the hierarchical structure of sentences by isolating and identifying the constituents. While the idea of breaking down sentences into smaller components can be traced back to early psychological and linguistic theories, ICA as a formal method was developed in the early 20th century. It was influenced by Wilhelm Wundt's psychological theories of sentence structure but was later refined and formalized within the framework of structural linguistics by Leonard Bloomfield. The method gained traction in the distributionalist tradition through the work of Zellig Harris and Charles F. Hockett, who expanded and applied it to sentence analysis. Additionally, ICA was further explored within the context of glossematics by Knud Togeby. These contributions helped ICA become a central tool in syntactic analysis, focusing on the hierarchical relationships between sentence constituents.
In syntax, shifting occurs when two or more constituents appearing on the same side of their common head exchange positions in a sense to obtain non-canonical order. The most widely acknowledged type of shifting is heavy NP shift, but shifting involving a heavy NP is just one manifestation of the shifting mechanism. Shifting occurs in most if not all European languages, and it may in fact be possible in all natural languages including sign languages. Shifting is not inversion, and inversion is not shifting, but the two mechanisms are similar insofar as they are both present in languages like English that have relatively strict word order. The theoretical analysis of shifting varies in part depending on the theory of sentence structure that one adopts. If one assumes relatively flat structures, shifting does not result in a discontinuity. Shifting is often motivated by the relative weight of the constituents involved. The weight of a constituent is determined by a number of factors: e.g., number of words, contrastive focus, and semantic content.
In linguistics, a discontinuity occurs when a given word or phrase is separated from another word or phrase that it modifies in such a manner that a direct connection cannot be established between the two without incurring crossing lines in the tree structure. The terminology that is employed to denote discontinuities varies depending on the theory of syntax at hand. The terms discontinuous constituent, displacement, long distance dependency, unbounded dependency, and projectivity violation are largely synonymous with the term discontinuity. There are various types of discontinuities, the most prominent and widely studied of these being topicalization, wh-fronting, scrambling, and extraposition.
In formal syntax, a node is a point in a tree diagram or syntactic tree that can be assigned a syntactic category label.