Endocentric and exocentric

Last updated

In theoretical linguistics, a distinction is made between endocentric and exocentric constructions. A grammatical construction (for instance, a phrase or compound) is said to be endocentric if it fulfils the same linguistic function as one of its parts, and exocentric if it does not. [1] The distinction reaches back at least to Bloomfield's work of the 1930s, [2] who based it on terms by Pāṇini and Patañjali in Sanskrit grammar. [3] Such a distinction is possible only in phrase structure grammars (constituency grammars), since in dependency grammars all constructions are necessarily endocentric. [4]

Contents

Endocentric construction

An endocentric construction consists of an obligatory head and one or more dependents, whose presence serves to modify the meaning of the head. For example:

  1. [NP [A big] [Nhouse]]
  2. [VP [Vsing] [N songs]]
  3. [AP [Adv very] [Along]]

These phrases are indisputably endocentric. They are endocentric because the one word in each case carries the bulk of the semantic content and determines the grammatical category to which the whole constituent will be assigned. The phrase big house is a noun phrase in line with its part house, which is a noun. Similarly, sing songs is a verb phrase in line with its part sing, which is a verb. The same is true of very long; it is an adjective phrase in line with its part long, which is an adjective. In more formal terms, the distribution of an endocentric construction is functionally equivalent, or approaching equivalence, to one of its parts, which serves as the center, or head, of the whole. An endocentric construction is also known as a headed construction, where the head is contained "inside" the construction.

Exocentric construction

An exocentric construction consists of two or more parts, whereby the one or the other of the parts cannot be viewed as providing the bulk of the semantic content of the whole. Further, the syntactic distribution of the whole cannot be viewed as being determined by the one or the other of the parts. The classic instance of an exocentric construction is the sentence (in a phrase structure grammar). [5] The traditional binary division [6] of the sentence (S) into a subject noun phrase (NP) and a predicate verb phrase (VP) was exocentric:

Hannibal destroyed Rome. - Sentence (S)

Since the whole is unlike either of its parts, it is exocentric. In other words, since the whole is neither a noun (N) like Hannibal nor a verb phrase (VP) like destroyed Rome but rather a sentence (S), it is exocentric. With the advent of X-bar theory in Transformational Grammar in the 1970s, this traditional exocentric division was largely abandoned and replaced by an endocentric analysis, whereby the sentence is viewed as an inflection phrase (IP), which is essentially a projection of the verb (a fact that makes the sentence a big VP in a sense). Thus, with the advent of X-bar theory, the endocentric vs. exocentric distinction started to become less important in transformational theories of syntax, for without the concept of exocentricity, the notion of endocentricity was becoming vacuous.

By contrast, in constraint-based syntactic theories, such as Lexical Functional Grammar (LFG), exocentric constructions are still widely used, but with a different role. Exocentricity is used in the treatment of non-configurational languages. As constraint-based models such as LFG do not represent a "deep structure" at which non-configurational languages can be treated as configurational, the exocentric S is used to formally represent the flat structure inherent in a non-configurational language. Hence, in a constraints-based analysis of Warlpiri, an exocentric structure follows the auxiliary, dominating all of the verb, arguments and adjuncts which are not raised to the specifier position of the IP:

[IP [NP Ngarrka-ngku][AUX ka][S [NP wawirri][V panti-rni]]]
'The man is spearing the kangaroo'

In addition, in theories of morphology, the distinction remains, since certain compounds seem to require an exocentric analysis, e.g. have-not in Bill is a have-not. For a class of compounds described as exocentric, see bahuvrihi.

The distinction in dependency grammars

The endo- vs. exocentric distinction is possible in phrase structure grammars (= constituency grammars), since they are constituency-based. The distinction is hardly present in dependency grammars, since they are dependency-based. In other words, dependency-based structures are necessarily endocentric, i.e. they are necessarily headed structures. Dependency grammars by definition were much less capable of acknowledging the types of divisions that constituency enables. Acknowledging exocentric structure necessitates that one posit more nodes in the syntactic (or morphological) structure than one has actual words or morphs in the phrase or sentence at hand. What this means is that a significant tradition in the study of syntax and grammar has been incapable from the start of acknowledging the endo- vs. exocentric distinction, a fact that has generated confusion about what should count as an endo- or exocentric structure.

Representing endo- and exocentric structures

Theories of syntax (and morphology) represent endocentric and exocentric structures using tree diagrams and specific labeling conventions. The distinction is illustrated here using the following trees. The first three trees show the distinction in a constituency-based grammar, and the second two trees show the same structures in a dependency-based grammar:

E-endo-01.jpg

The upper two trees on the left are endocentric since each time, one of the parts, i.e. the head, projects its category status up to the mother node. The upper tree on the right, in contrast, is exocentric, because neither of the parts projects its category status up to the mother node; Z is a category distinct from X or Y. The two dependency trees show the manner in which dependency-based structures are inherently endocentric. Since the number of nodes in the tree structure is necessarily equal to the number of elements (e.g. words) in the string, there is no way to assign the whole (i.e. XY) a category status that is distinct from both X and Y.

Traditional phrase structure trees are mostly endocentric, although the initial binary division of the clause is exocentric (S → NP VP), as mentioned above, e.g.

E-endo-02.jpg

This tree structure contains four divisions, whereby only one of these division is exocentric (the highest one). The other three divisions are endocentric because the mother node has the same basic category status as one of its daughters. The one exocentric division disappears in the corresponding dependency tree:

E-endo-03.jpg

Dependency positions the finite verb as the root of the entire tree, which means the initial exocentric division is impossible. This tree is entirely endocentric.

In languages

Chinese

The Chinese language is known for having rich compounds. [7] Linguists often classify compound verbs in Chinese into five types: Subject-Predicate 主謂結構 (SP), Verb-Object 述賓結構 (VO), Verb-Complement 述補結構 (VC), Coordinative 並列結構 (VV), and Endocentric 偏正結構. [8] [9] The Coordinative, Verb-Complement, and Endocentric types are also known as Parallel, Verb-Resultative, and Modifier-Head, respectively. [10]

Below are a few examples of the exocentric compounds in Chinese. [11] [12]

ExampleInternal StructureExplanation
大小dà-xiǎoA-A → Nbig + small → size
好歹hǎo-dǎiA-A → Advgood + bad → anyhow
開關kāi-guānV-V → Nopen + close → switch
保守bǎo-shǒuV-V → Akeep + defend → conservative
物色wù-sèN-N → Vitem + color → choose from
矛盾máo-dùnN-N → Aspear + shield → contradictory

Warlpiri

The Warlpiri language is widely held as the canonical example of a non-configurational language. [13] As such, Warlpiri sentences exhibit exceptionally flat surface structure. If a non-derivational approach is taken to syntactic structure, this can best be formalised with exocentric S dominated by the auxiliary in I. Thus, an example analysis of the constituent structure of the Warlpiri sentence:

Ngarrka-ngku

man-ERG

ka

AUX

wawirri

kangaroo.ABS

panti-rni

spear-NPAST

Ngarrka-ngku ka wawirri panti-rni

man-ERG AUX kangaroo.ABS spear-NPAST

'the man is spearing the kangaroo'

would be as follows:

Constituent structure tree diagram for Warlpiri sentence "the man is spearing the kangaroo" Warlpiri c-structure.jpg
Constituent structure tree diagram for Warlpiri sentence "the man is spearing the kangaroo"

Where S is a non-projected exocentric structure which dominates both heads and phrases with equal weight. The elements in spec of IP and under S can be freely moved and switch places, as position in c-structure, except for I, plays a pragmatic rather than syntactic role in a constraints-based analysis of Warlpiri sentence structure.

A note about coordinate structures

While exocentric structures have largely disappeared from most theoretical analyses of standard sentence structure, many theories of syntax still assume (something like) exocentric divisions for coordinate structures, e.g.

[Sam] and [Larry] arrived.
She [laughed] and [cried].
[Should I] or [should I not] go to that conference?

The brackets each time mark the conjuncts of a coordinate structure, whereby this coordinate structure includes the material appearing between the left-most bracket and the right-most bracket; the coordinator is positioned between the conjuncts. Coordinate structures like these do not lend themselves to an endocentric analysis in any clear way, nor to an exocentric analysis. One might argue that the coordinator is the head of the coordinate structure, which would make it endocentric. This argument would have to ignore the numerous occurrences of coordinate structures that lack a coordinator (asyndeton), however. One might therefore argue instead that coordinate structures like these are multi-headed, each conjunct being or containing a head. The difficulty with this argument, however, is that the traditional endocentric vs. exocentric distinction did not foresee the existence of multi-headed structures, which means that it did not provide a guideline for deciding whether a multi-headed structure should be viewed as endo- or exocentric. Coordinate structures thus remain a problem area for the endo- vs. exocentric distinction in general.

Notes

  1. Matthews (1981:147) provides an insightful discussion of the endo- vs. exocentric distinction. See Falk (2001:43ff., 49ff.) as well.
  2. See Bloomfield (1933), 194–196 and 235–237.
  3. Wujastyk (1982).
  4. Concerning the lack of exocentric structures in dependency grammar, see Osborne et al. (2019: 48-50).
  5. Concerning the status of S as an exocentric construction, see Emonds (1976:15).
  6. See for example Chomsky (1957).
  7. Arcodia, Giorgio Francesco. (2007). Chinese: A language of compound words? In F. Montermini, G. Boyé, & N. Hathout (Eds.), Selected Proceedings of the 5th Décembrettes: Morphology in Toulouse (pp. 79-90). Somerville, MA: Cascadilla Proceedings Project.
  8. Li, D.-J. & Cheng, M.-Z. (2008). A Practical Chinese Grammar for Foreigners (Rev. ed.). Beijing: Beijing Language and Culture University Press.
  9. Chang, S.-M. & Tang, T.-C. (2009). On the Study of Compounds: A Contrastive Analysis of Chinese, English and Japanese. Journal of Taiwanese Languages and Literature, 3, 179-213.
  10. Liao, W.-W. R. (2014). Morphology. In C.-T. Huang, Y.-H. Li, & A. Simpson (Eds), The Handbook of Chinese Linguistics (pp. 3-25). Malden, MA: Wiley Blackwell.
  11. Zhang, N. N. (2007). Root merger in Chinese compounds. Studia Linguistica, 61(2), 170-184.
  12. Scalise, S., Fábregas, A., & Forza, F. (2009). Exocentricity in Compounding. 言語研究 (Gengo Kenkyu), 135, 49-84.
  13. Hale, K. (1983). "Warlpiri and the grammar of non-configurational languages". Natural Language and Linguistic Theory. 2 (1): 39–76.

Related Research Articles

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning. There are numerous approaches to syntax which differ in their central assumptions and goals.

A syntactic category is a syntactic unit that theories of syntax assume. Word classes, largely corresponding to traditional parts of speech are syntactic categories. In phrase structure grammars, the phrasal categories are also syntactic categories. Dependency grammars, however, do not acknowledge phrasal categories.

In syntax and grammar, a phrase is a group of words which act together as a grammatical unit. For instance, the English expression "the very happy squirrel" is a noun phrase which contains the adjective phrase "very happy". Phrases can consist of a single word or a complete sentence. In theoretical linguistics, phrases are often analyzed as units of syntactic structure such as a constituent.

Phrase structure rules are a type of rewrite rule used to describe a given language's syntax and are closely associated with the early stages of transformational grammar, proposed by Noam Chomsky in 1957. They are used to break down a natural language sentence into its constituent parts, also known as syntactic categories, including both lexical categories and phrasal categories. A grammar that uses phrase structure rules is a type of phrase structure grammar. Phrase structure rules as they are commonly employed operate according to the constituency relation, and a grammar that employs phrase structure rules is therefore a constituency grammar; as such, it stands in contrast to dependency grammars, which are based on the dependency relation.

In linguistics, a noun phrase, or nominal (phrase), is a phrase that has a noun or pronoun as its head or performs the same grammatical function as a noun. Noun phrases are very common cross-linguistically, and they may be the most frequently occurring phrase type.

Parse tree

A parse tree or parsing tree or derivation tree or concrete syntax tree is an ordered, rooted tree that represents the syntactic structure of a string according to some context-free grammar. The term parse tree itself is used primarily in computational linguistics; in theoretical syntax, the term syntax tree is more common.

In linguistics, X-bar theory is a model of phrase-structure grammar and a theory of syntactic category formation that was first proposed by Noam Chomsky in 1970 and further developed by Jackendoff, along the lines of the theory of generative grammar put forth in the 1950s by Chomsky. It attempts to capture the structure of phrasal categories with a single uniform structure called the X-bar schema, basing itself on the assumption that any phrase in natural language is an XP that is headed by a given syntactic category X. It played a significant role in resolving issues that phrase structure rules had, representative of which is the proliferation of grammatical rules, which is against the thesis of generative grammar.

In linguistics, a verb phrase (VP) is a syntactic unit composed of at least one verb and its dependents—objects, complements and other modifiers—but not always including the subject. Thus in the sentence A fat man put the money quickly in the box, the words put the money quickly in the box are a verb phrase; it consists of the verb put and its dependents, but not the subject a fat man. A verb phrase is similar to what is considered a predicate in more traditional grammars.

In linguistics, the head or nucleus of a phrase is the word that determines the syntactic category of that phrase. For example, the head of the noun phrase boiling hot water is the noun water. Analogously, the head of a compound is the stem that determines the semantic category of that compound. For example, the head of the compound noun handbag is bag, since a handbag is a bag, not a hand. The other elements of the phrase or compound modify the head, and are therefore the head's dependents. Headed phrases and compounds are called endocentric, whereas exocentric ("headless") phrases and compounds lack a clear head. Heads are crucial to establishing the direction of branching. Head-initial phrases are right-branching, head-final phrases are left-branching, and head-medial phrases combine left- and right-branching.

In linguistics, branching refers to the shape of the parse trees that represent the structure of sentences. Assuming that the language is being written or transcribed from left to right, parse trees that grow down and to the right are right-branching, and parse trees that grow down and to the left are left-branching. The direction of branching reflects the position of heads in phrases, and in this regard, right-branching structures are head-initial, whereas left-branching structures are head-final. English has both right-branching (head-initial) and left-branching (head-final) structures, although it is more right-branching than left-branching. Some languages such as Japanese and Turkish are almost fully left-branching (head-final). Some languages are mostly right-branching (head-initial).

Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.

The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue. Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining trait of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.

A sentence diagram is a pictorial representation of the grammatical structure of a sentence. The term "sentence diagram" is used more when teaching written language, where sentences are diagrammed. The model shows the relations between words and the nature of sentence structure and can be used as a tool to help recognize which potential sentences are actual sentences.

In generative grammar, non-configurational languages are languages characterized by a flat phrase structure, which allows syntactically discontinuous expressions, and a relatively free word order.

Topicalization is a mechanism of syntax that establishes an expression as the sentence or clause topic by having it appear at the front of the sentence or clause. This involves a phrasal movement of determiners, prepositions, and verbs to sentence-initial position. Topicalization often results in a discontinuity and is thus one of a number of established discontinuity types, the other three being wh-fronting, scrambling, and extraposition. Topicalization is also used as a constituency test; an expression that can be topicalized is deemed a constituent. The topicalization of arguments in English is rare, whereas circumstantial adjuncts are often topicalized. Most languages allow topicalization, and in some languages, topicalization occurs much more frequently and/or in a much less marked manner than in English. Topicalization in English has also received attention in the pragmatics literature.

Grammatical relation

In linguistics, grammatical relations are functional relationships between constituents in a clause. The standard examples of grammatical functions from traditional grammar are subject, direct object, and indirect object. In recent times, the syntactic functions, typified by the traditional categories of subject and object, have assumed an important role in linguistic theorizing, within a variety of approaches ranging from generative grammar to functional and cognitive theories. Many modern theories of grammar are likely to acknowledge numerous further types of grammatical relations. The role of grammatical relations in theories of grammar is greatest in dependency grammars, which tend to posit dozens of distinct grammatical relations. Every head-dependent dependency bears a grammatical function.

In linguistics, coordination is a complex syntactic structure that links together two or more elements; these elements are called conjuncts or conjoins. The presence of coordination is often signaled by the appearance of a coordinator, e.g. and, or, but. The totality of coordinator(s) and conjuncts forming an instance of coordination is called a coordinate structure. The unique properties of coordinate structures have motivated theoretical syntax to draw a broad distinction between coordination and subordination. It is also one of the many constituency tests in linguistics. Coordination is one of the most studied fields in theoretical syntax, but despite decades of intensive examination, theoretical accounts differ significantly and there is no consensus on the best analysis.

In linguistics, subordination is a principle of the hierarchical organization of linguistic units. While the principle is applicable in semantics, syntax, morphology, and phonology, most work in linguistics employs the term "subordination" in the context of syntax, and that is the context in which it is considered here. The syntactic units of sentences are often either subordinate or coordinate to each other. Hence an understanding of subordination is promoted by an understanding of coordination, and vice versa.

Merge is one of the basic operations in the Minimalist Program, a leading approach to generative syntax, when two syntactic objects are combined to form a new syntactic unit. Merge also has the property of recursion in that it may apply to its own output: the objects combined by Merge are either lexical items or sets that were themselves formed by Merge. This recursive property of Merge has been claimed to be a fundamental characteristic that distinguishes language from other cognitive faculties. As Noam Chomsky (1999) puts it, Merge is "an indispensable operation of a recursive system ... which takes two syntactic objects A and B and forms the new object G={A,B}" (p. 2).

Subject–verb inversion in English is a type of inversion where the subject and verb switch their canonical order of appearance so that the subject follows the verb(s), e.g. A lamp stood beside the bedBeside the bed stood a lamp. Subject–verb inversion is distinct from subject–auxiliary inversion because the verb involved is not an auxiliary verb.

References

See also