Branching (linguistics)

Last updated

In linguistics, branching refers to the shape of the parse trees that represent the structure of sentences. [1] Assuming that the language is being written or transcribed from left to right, parse trees that grow down and to the right are right-branching, and parse trees that grow down and to the left are left-branching. The direction of branching reflects the position of heads in phrases, and in this regard, right-branching structures are head-initial , whereas left-branching structures are head-final. [2] English has both right-branching (head-initial) and left-branching (head-final) structures, although it is more right-branching than left-branching. [3] Some languages such as Japanese and Turkish are almost fully left-branching (head-final). Some languages are mostly right-branching (head-initial).

Contents

Examples

Languages typically construct phrases with a head word (or nucleus) and zero or more dependents (modifiers). The following phrases show the phrase heads in bold.

Examples of left-branching phrases (= head-final phrases):

the house- Noun phrase (NP)
very happy- Adjective phrase (AP)
too slowly- Adverb phrase (AdvP)

Examples of right-branching phrases (= head-initial phrases):

laugh loudly - Verb phrase (VP)
with luck - Prepositional phrase (PP)
that it happened - Subordinator phrase (SP = subordinate clause)

Examples of phrases that contain both left- and right-branching (= head-medial phrases):

the house there - Noun phrase (NP)
very happy with it - Adjective phrase (AP)
only laugh loudly - Verb phrase (VP)

Concerning phrases such as the house and the house there, this article assumes the traditional NP analysis, meaning that the noun is deemed to be head over the determiner. On a DP-analysis (determiner phrase), the phrase the house would be right-branching instead of left-branching.

Tree structures

Left- and right-branching structures are illustrated with the trees that follow. Each example appears twice, once according to a constituency-based analysis associated with a phrase structure grammar [4] and once according to a dependency-based analysis associated with a dependency grammar. [5] The first group of trees illustrate left-branching:

Branching1.jpg

The upper row shows the constituency-based structures, and the lower row the dependency-based structures. In the constituency-based structures, left-branching is present (but not really visible) in so far as the non-head daughter is to the left of the head. In the corresponding dependency-based structures in the lower row, the left-branching is clear; the dependent appears to the left of its head, the branch extending down to the left. The following structures demonstrate right-branching:

Branching2.jpg

The upper row again shows the constituency-based structures, and the lower row the dependency-based structures. The constituency-based structures are right-branching insofar as the non-head daughter is to the right of the head. This right-branching is completely visible in the lower row of dependency-based structures, where the branch extends down to the right. The (c)-examples contain one instance of right-branching (the upper branch) and one instance of left-branching (the lower branch). The following trees illustrate phrases that combine both types of branching:

Branching3.jpg

The combination of left- and right-branching is now completely visible in both the constituency- and dependency-based trees. The head appears in a medial position, which means that the phrase combines both types of branching. Note that the (b)-trees also contain a PP phrase that is an instance of pure right-branching.

Full trees

The nature of branching is most visible with full trees. The following trees have been chosen to illustrate the extent to which a structure can be entirely left- or entirely right-branching. The following sentence is completely left-branching. The constituency-based trees are on the left, and the dependency-based trees are on the right: [6]

Branching4.jpg

The category Po (= possessive) is used to label possessive 's. The following sentence is completely right-branching:

Branching5.jpg

Most structures in English are, however, not completely left- or completely right-branching, but rather they combine both. The following trees illustrate what can be seen as a stereotypical combination of left- and right-branching in English:

Branching6.jpg

Determiners (e.g. the) always and subjects (e.g. the child) usually appear on left branches in English, but infinitival verbs (e.g. try, eat) and the verb particle to usually appear on right branches. In the big picture, right-branching structures tend to outnumber the left-branching structures in English, which means that trees usually grow down to the right.

X-bar schema

The X-bar schema [7] combines left- and right-branching. The standard X-bar schema has the following structure:

Branching7.jpg

This structure is both left- and right branching. It is left-branching insofar as the bar-level projection of the head (X') follows the specifier, but it is right-branching insofar as the actual head (X0) precedes the complement. Despite these conflicting traits, most standard X-bar structures (in English) are more right-branching than left-branching because specifiers tend to be less complex (i.e. fewer words) than complements.

Binary vs. n-ary branching

Much work in Government and Binding Theory (GB), the Minimalist Program (MP), and Lexical Functional Grammar (LFG) assumes all branching to be binary. [8] Other theories (both constituency- and dependency-based ones), e.g. early Transformational Grammar, Head-Driven Phrase Structure Grammar, Meaning-Text Theory, Word Grammar, etc. allow for n-ary branching. This distinction can have a profound impact on the overall nature of the theory of syntax. The two main possibilities in a phrase structure grammar are illustrated with the following trees:

Branching8.jpg

The binary branching on the left is closely associated with the structures of GB, MP, and LFG, and it is similar to what the X-bar schema assumes. The n-ary branching structure on the right is a more traditional approach to branching. One can muster arguments for both approaches. For instance, the critics of the strictly binary branching structures charge that the strict binarity is motivated more by a desire for theoretical purity than by empirical observation.

Strictly binary branching structures increase the amount of syntactic structure (number of nodes) to the upper limit of what is possible, whereas flatter n-ary branching tends to restrict the amount of structure that the theory can assume. Worth noting in this area is that the more layered the syntactic structures are, the more discontinuities can occur, which means the component of the theory that addresses discontinuities must play a greater role. Given the flatter structures associated with n-ary branching, certain phenomena (e.g. inversion and shifting) do not result in discontinuities, a fact that reduces the role that the component for discontinuities must play in the theory.

Tendencies

As stated above, the main branching trait for a language is just a tendency and it often shows exceptions. Spanish, for example, while overwhelmingly right-branching, puts numeral modifiers before nouns and, in certain cases, objects before verbs. Languages like English or Swedish, though regarded as being right-branching because the main verbs precede direct objects, place adjectives and numerals before their nouns. On the contrary, Northwest Caucasian languages such as Ubykh or Abkhaz are strongly left-branching but put adjectives after nouns.

Some languages, however, are exclusively left-branching or right-branching. Japanese and most other languages of northeastern Asia and the Indian subcontinent, as well as the Turkic languages, are practically a model for rigidly left-branching languages. The Mon–Khmer and Austronesian languages of southeast Asia and many African languages come close to rigidly right-branching, with numerals as well as adjectives following their nouns and with degree words like very, too, extremely, and quite following the adjectives they modify.

See also

Notes

  1. For a comprehensive discussion of branching, see Berg (2009:34ff.).
  2. Concerning the correlation between head position and branching direction, see Payne (2006:194).
  3. The tendency for English to be right-branching is widely acknowledged, e.g. van Riemsdijk and Williams (1986:211).
  4. Phrase structure grammars are those grammars that follow in the tradition of Chomsky (1957).
  5. Dependency grammars are those grammars that follow in the tradition of Tesnière (1959)
  6. For similar, constituency-based examples of purely left- and then purely right-branching structures, see Fowler (1971:125f.).
  7. See Chomsky (1970) and Jackendoff (1977) for the discussions of the X-bar schema.
  8. The work of Larson (1988) and Kayne (1981, 1994) were influential in establishing strictly binary branching structures in the tradition of Chomskyan syntax.

Related Research Articles

In linguistics, syntax is the set of rules, principles, and processes that govern the structure of sentences in a given language, usually including word order. The term syntax is also used to refer to the study of such principles and processes. The goal of many syntacticians is to discover the syntactic rules common to all languages.

A syntactic category is a syntactic unit that theories of syntax assume. Word classes, largely corresponding to traditional parts of speech are syntactic categories. In phrase structure grammars, the phrasal categories are also syntactic categories. Dependency grammars, however, do not acknowledge phrasal categories.

In everyday speech, a phrase is any group of words, often carrying a special idiomatic meaning; in this sense it is synonymous with expression and phraseme. In linguistic analysis, a phrase is a group of words that functions as a constituent in the syntax of a sentence, a single unit within a grammatical hierarchy. A phrase typically appears within a clause, but it is possible also for a phrase to be a clause or to contain a clause within it. There are also types of phrases like noun phrase and prepositional phrase.

Phrase structure rules are a type of rewrite rule used to describe a given language's syntax and are closely associated with the early stages of transformational grammar, proposed by Noam Chomsky in 1957. They are used to break down a natural language sentence into its constituent parts, also known as syntactic categories, including both lexical categories and phrasal categories. A grammar that uses phrase structure rules is a type of phrase structure grammar. Phrase structure rules as they are commonly employed operate according to the constituency relation, and a grammar that employs phrase structure rules is therefore a constituency grammar; as such, it stands in contrast to dependency grammars, which are based on the dependency relation.

A noun phrase, or nominal (phrase), is a phrase that has a noun as its head or performs the same grammatical function as a noun. Noun phrases are very common cross-linguistically, and they may be the most frequently occurring phrase type.

Parse tree

A parse tree or parsing tree or derivation tree or concrete syntax tree is an ordered, rooted tree that represents the syntactic structure of a string according to some context-free grammar. The term parse tree itself is used primarily in computational linguistics; in theoretical syntax, the term syntax tree is more common.

In linguistics, X-bar theory is a theory of syntactic category formation. It embodies two independent claims: one, that phrases may contain intermediate constituents projected from a head X; and two, that this system of projected constituency may be common to more than one category.

In linguistics, the minimalist program (MP) is a major line of inquiry that has been developing inside generative grammar since the early 1990s, starting with a 1993 paper by Noam Chomsky.

In linguistics, the head or nucleus of a phrase is the word that determines the syntactic category of that phrase. For example, the head of the noun phrase boiling hot water is the noun water. Analogously, the head of a compound is the stem that determines the semantic category of that compound. For example, the head of the compound noun handbag is bag, since a handbag is a bag, not a hand. The other elements of the phrase or compound modify the head, and are therefore the head's dependents. Headed phrases and compounds are called endocentric, whereas exocentric ("headless") phrases and compounds lack a clear head. Heads are crucial to establishing the direction of branching. Head-initial phrases are right-branching, head-final phrases are left-branching, and head-medial phrases combine left- and right-branching.

Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.

The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue. Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining trait of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.

A sentence diagram is a pictorial representation of the grammatical structure of a sentence. The term "sentence diagram" is used more when teaching written language, where sentences are diagrammed. The model shows the relations between words and the nature of sentence structure and can be used as a tool to help recognize which potential sentences are actual sentences.

In theoretical linguistics, a distinction is made between endocentric and exocentric constructions. A grammatical construction is said to be endocentric if it fulfils the same linguistic function as one of its parts, and exocentric if it does not. The distinction reaches back at least to Bloomfield's work of the 1930s. Such a distinction is possible only in phrase structure grammars, since in dependency grammars all constructions are necessarily endocentric.

Lucien Tesnière French linguist

Lucien Tesnière was a prominent and influential French linguist. He was born in Mont-Saint-Aignan on May 13, 1893. As a maître de conférences in Strasbourg (1924), and later professor in Montpellier (1937), he published many papers and books on Slavic languages. However, his importance in the history of linguistics is based mainly on his development of an approach to the syntax of natural languages that would become known as dependency grammar. He presented his theory in his book Éléments de syntaxe structurale, published posthumously in 1959. In the book he proposes a sophisticated formalization of syntactic structures, supported by many examples from a diversity of languages. Tesnière died in Montpellier on December 6, 1954.

In linguistics, antisymmetry is a theory of syntactic linearization presented in Richard S. Kayne's 1994 monograph The Antisymmetry of Syntax. The crux of this theory is that hierarchical structure in natural language maps universally onto a particular surface linearization, namely specifier-head-complement branching order. The theory derives a version of X-bar theory. Kayne hypothesizes that all phrases whose surface order is not specifier-head-complement have undergone syntactic movements that disrupt this underlying order. Subsequently, there have also been attempts at deriving specifier-complement-head as the basic word order.

Merge is one of the basic operations in the Minimalist Program, a leading approach to generative syntax, when two syntactic objects are combined to form a new syntactic unit. Merge also has the property of recursion in that it may apply to its own output: the objects combined by Merge are either lexical items or sets that were themselves formed by Merge. This recursive property of Merge has been claimed to be a fundamental characteristic that distinguishes language from other cognitive faculties. As Noam Chomsky (1999) puts it, Merge is "an indispensable operation of a recursive system ... which takes two syntactic objects A and B and forms the new object G={A,B}" (p. 2).

Syntactic movement is the means by which some theories of syntax address discontinuities. Movement was first postulated by structuralist linguists who expressed it in terms of discontinuous constituents or displacement. Some constituents appear to have been displaced from the position in which they receive important features of interpretation. The concept of movement is controversial and is associated with so-called transformational or derivational theories of syntax. Representational theories, in contrast, reject the notion of movement and often instead address discontinuities in terms of feature passing or persistent structural identities.

Scrambling is a syntactic phenomenon wherein sentences can be formulated using a variety of different word orders without any change in meaning. Scrambling often results in a discontinuity since the scrambled expression can end up at a distance from its head. Scrambling does not occur in English, but it is frequent in languages with freer word order, such as German, Russian, Persian and Turkic languages. The term was coined by Haj Ross in his 1967 dissertation and is widely used in present work, particularly with the generative tradition.

In syntax, shifting occurs when two or more constituents appearing on the same side of their common head exchange positions in a sense to obtain non-canonical order. The most widely acknowledged type of shifting is heavy NP shift, but shifting involving a heavy NP is just one manifestation of the shifting mechanism. Shifting occurs in most if not all European languages, and it may in fact be possible in all natural languages including sign languages.. Shifting is not inversion, and inversion is not shifting, but the two mechanisms are similar insofar as they are both present in languages like English that have relatively strict word order. The theoretical analysis of shifting varies in part depending on the theory of sentence structure that one adopts. If one assumes relatively flat structures, shifting does not result in a discontinuity. Shifting is often motivated by the relative weight of the constituents involved. The weight of a constituent is determined by a number of factors: e.g., number of words, contrastive focus, and semantic content.

In linguistics, a discontinuity occurs when a given word or phrase is separated from another word or phrase that it modifies in such a manner that a direct connection cannot be established between the two without incurring crossing lines in the tree structure. The terminology that is employed to denote discontinuities varies depending on the theory of syntax at hand. The terms discontinuous constituent, displacement, long distance dependency, unbounded dependency, and projectivity violation are largely synonymous with the term discontinuity. There are various types of discontinuities, the most prominent and widely studied of these being topicalization, wh-fronting, scrambling, and extraposition.

References