Selection (linguistics)

Last updated

In linguistics, selection denotes the ability of predicates to determine the semantic content of their arguments. [1] Predicates select their arguments, which means they limit the semantic content of their arguments. One sometimes draws a distinction between types of selection; one acknowledges both s(emantic)-selection and c(ategory)-selection. Selection in general stands in contrast to subcategorization: [2] predicates both select and subcategorize for their complement arguments, whereas they only select their subject arguments. Selection is a semantic concept, whereas subcategorization is a syntactic one. [3] Selection is closely related to valency, a term used in other grammars than the Chomskian generative grammar, for a similar phenomenon.

Contents

Examples

The following pairs of sentences will illustrate the concept of selection:

a. The plant is wilting.
b. #The building is wilting. - The argument the building violates the selectional restrictions of the predicate is wilting.
a. Sam drank a coffee.
b. #Sam drank a car. - The argument a car contradicts the selectional restrictions of the predicate drank.

The # indicates semantic deviance. The predicate is wilting selects a subject argument that is a plant or is plant-like. Similarly, the predicate drank selects an object argument that is a liquid or is liquid-like. A building cannot normally be understood as wilting, just as a car cannot normally be interpreted as a liquid. The b-sentences are possible only given an unusual context that establishes appropriate metaphorical meaning. The deviance of the b-sentences is addressed in terms of selection. The selectional restrictions of the predicates is wilting and drank are violated.

When a mismatch between a selector and a selected element triggers reinterpretation of the meaning of those elements, that process is referred to as coercion. [4]

S-selection vs. c-selection

One sometimes encounters the terms s(emantic)-selection and c(ategory)-selection. [5] The concept of c-selection overlaps to an extent with subcategorization. Predicates c-select the syntactic category of their complement arguments - e.g. noun (phrase), verb (phrase), adjective (phrase), etc. - i.e. they determine the syntactic category of their complements. In contrast, predicates s-select the semantic content of their arguments. Thus s-selection is a semantic concept, whereas c-selection is a syntactic one. When the term selection or selectional restrictions appears alone without the c- or s-, s-selection is usually understood. [6] [7]

The b-sentences above do not contain violations of the c-selectional restrictions of the predicates is wilting and drank; they are, rather, well-formed from a syntactic point of view (hence #, not *), for the arguments the building and a car satisfy the c-selectional restrictions of their respective predicates, these restrictions requiring their arguments to be nouns or noun phrases. Just the s-selectional restrictions of the predicates is wilting and drank are violated in the b-sentences.

Selectional constraints or selectional preferences describe the degree of s-selection, in contrast to selectional restrictions which treat s-selection as a binary, yes or no. [8] Selectional preferences have often been used as a source of linguistic information in natural language processing applications. [9] Thematic fit is a measure of how much a particular word in a particular role (like subject or direct object) matches the selectional preference of a particular predicate. For example, the word cake has a high thematic fit as a direct object for cut. [10]

C-selection vs. subcategorization

The concepts of c-selection and subcategorization overlap in meaning and use to a significant degree. [11] If there is a difference between these concepts, it resides with the status of the subject argument. Traditionally, predicates are interpreted as NOT subcategorizing for their subject argument because the subject argument appears outside of the minimal VP containing the predicate. [12] Predicates do, however, c-select their subject arguments, e.g.

Fred eats beans.

The predicate eats c-selects both its subject argument Fred and its object argument beans, but as far as subcategorization is concerned, eats subcategorizes for its object argument beans only. This difference between c-selection and subcategorization depends crucially on the understanding of subcategorization. An approach to subcategorization that sees predicates as subcategorizing for their subject arguments as well as for their object arguments will draw no distinction between c-selection and subcategorization; the two concepts are synonymous for such approaches.

Thematic relations

Selection can be closely associated with thematic relations (e.g. agent, patient, theme, goal, etc.). [13] By limiting the semantic content of their arguments, predicates are determining the thematic relations/roles that their arguments bear.

Theories

Several linguistic theories make explicit use of selection. These include:

Notes

  1. For discussions of selection in general, see Chomsky (1965), Horrocks (1986:35f.), van Riemsdijk and Williams (1986:130), Cowper (1992:58), Napoli (1993:260ff.), Carnie (2007:220-221).
  2. See Fowler (1971:58) concerning the distinction between selection and subcategorization.
  3. Resnik, P. (1993). Semantic classes and syntactic ambiguity . In HUMAN LANGUAGE TECHNOLOGY: Proceedings of a Workshop Held at Plainsboro, New Jersey, March 21-24, 1993, p.279, "selectional preference (..) a class of restrictions on co-occurrence that is orthogonal to syntactic constraints"
  4. Lauwers, P.; Willems, D. (2011). "Coercion: Definition and challenges, current approaches, and new trends". Linguistics. 49 (6): 1219–1235. doi:10.1515/ling.2011.034. hdl: 1854/LU-2046811 .
  5. Concerning the distinction between s-selection and c-selection, see for instance Ouhalla (1994:125), Lasnik (1999:21), and Fromkin et al. (2000:228ff.).
  6. For examples of selection used in the sense of "s-selection", see for instance Chisholm (1981:139), Brinton (2000:153), van Valin (2001:87).
  7. Haegeman and Guéron (1999:22f), however, mean c-selection when they write just selection.
  8. Resnik, Philip (October 1, 1996). "Selectional constraints: An information-theoretic model and its computational realization". Cognition. 61 (1–2): 127–159.
  9. Roberts, W., & Egg, M. (2014, October). A comparison of selectional preference models for automatic verb classification . In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (pp. 511-522).
  10. Sayeed, A., Greenberg, C., & Demberg, V. (2016, August). Thematic fit evaluation: an aspect of selectional preferences . In Proceedings of the 1st Workshop on Evaluating Vector-Space Representations for NLP (pp. 99-105).
  11. Concerning the overlap in meaning and use of the terms c-selection and subcategorization, see Fromkin (2000:230).
  12. See for instance Chomsky's (1965) original discussion of subcategorization.
  13. Concerning the connection between selection and thematic relations/roles, see Ouhalla (125).

Literature

Related Research Articles

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning. There are numerous approaches to syntax which differ in their central assumptions and goals.

In language, a clause is a constituent that links a semantic predicand and a semantic predicate. A typical clause consists of a subject and a syntactic predicate, the latter typically a verb phrase, a verb with any objects and other modifiers. However, the subject is sometimes not said or explicit, often the case in null-subject languages if the subject is retrievable from context, but it sometimes also occurs in other languages such as English.

Lexical semantics, as a subfield of linguistic semantics, is the study of word meanings. It includes the study of how words structure their meaning, how they act in grammar and compositionality, and the relationships between the distinct senses and uses of a word.

In generative grammar, a theta role or θ-role is the formal device for representing syntactic argument structure—the number and type of noun phrases—required syntactically by a particular verb. For example, the verb put requires three arguments.

Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.

In grammar and theoretical linguistics, government or rection refers to the relationship between a word and its dependents. One can discern between at least three concepts of government: the traditional notion of case government, the highly specialized definition of government in some generative models of syntax, and a much broader notion in dependency grammars.

In linguistics, valency or valence is the number and type of arguments controlled by a predicate, content verbs being typical predicates. Valency is related, though not identical, to subcategorization and transitivity, which count only object arguments – valency counts all arguments, including the subject. The linguistic meaning of valency derives from the definition of valency in chemistry. The valency metaphor appeared first in linguistics in Charles Sanders Peirce's essay "The Logic of Relatives" in 1897, and it then surfaced in the works of a number of linguists decades later in the late 1940s and 1950s. Lucien Tesnière is credited most with having established the valency concept in linguistics. A major authority on the valency of the English verbs is Allerton (1982), who made the important distinction between semantic and syntactic valency.

In generative grammar and related frameworks, a node in a parse tree c-commands its sister node and all of its sister's descendants. In these frameworks, c-command plays a central role in defining and constraining operations such as syntactic movement, binding, and scope. Tanya Reinhart introduced c-command in 1976 as a key component of her theory of anaphora. The term is short for "constituent command".

In grammar, a complement is a word, phrase, or clause that is necessary to complete the meaning of a given expression. Complements are often also arguments.

The term predicate is used in one of two ways in linguistics and its subfields. The first defines a predicate as everything in a standard declarative sentence except the subject, and the other views it as just the main content verb or associated predicative expression of a clause. Thus, by the first definition the predicate of the sentence Frank likes cake is likes cake. By the second definition, the predicate of the same sentence is just the content verb likes, whereby Frank and cake are the arguments of this predicate. Differences between these two definitions can lead to confusion.

Grammatical relation

In linguistics, grammatical relations are functional relationships between constituents in a clause. The standard examples of grammatical functions from traditional grammar are subject, direct object, and indirect object. In recent times, the syntactic functions, typified by the traditional categories of subject and object, have assumed an important role in linguistic theorizing, within a variety of approaches ranging from generative grammar to functional and cognitive theories. Many modern theories of grammar are likely to acknowledge numerous further types of grammatical relations. The role of grammatical relations in theories of grammar is greatest in dependency grammars, which tend to posit dozens of distinct grammatical relations. Every head-dependent dependency bears a grammatical function.

In linguistics, raising constructions involve the movement of an argument from an embedded or subordinate clause to a matrix or main clause; in other words, a raising predicate/verb appears with a syntactic argument that is not its semantic argument, but is rather the semantic argument of an embedded predicate. For example, in they seem to be trying, the predicand of trying is the subject of seem. Although English has raising constructions, not all languages do.

In linguistics, control is a construction in which the understood subject of a given predicate is determined by some expression in context. Stereotypical instances of control involve verbs. A superordinate verb "controls" the arguments of a subordinate, nonfinite verb. Control was intensively studied in the government and binding framework in the 1980s, and much of the terminology from that era is still used today. In the days of Transformational Grammar, control phenomena were discussed in terms of Equi-NP deletion. Control is often analyzed in terms of a null pronoun called PRO. Control is also related to raising, although there are important differences between control and raising. Most if not all languages have control constructions and these constructions tend to occur frequently.

In generative grammar and related approaches, the logical Form (LF) of a linguistic expression is the variant of its syntactic structure which undergoes semantic interpretation. It is distinguished from phonetic form, the structure which corresponds to a sentence's pronunciation. These separate representations are postulated in order to explain the ways in which an expression's meaning can be partially independent of its pronunciation, e.g. scope ambiguities.

In linguistics, an argument is an expression that helps complete the meaning of a predicate, the latter referring in this context to a main verb and its auxiliaries. In this regard, the complement is a closely related concept. Most predicates take one, two, or three arguments. A predicate and its arguments form a predicate-argument structure. The discussion of predicates and arguments is associated most with (content) verbs and noun phrases (NPs), although other syntactic categories can also be construed as predicates and as arguments. Arguments must be distinguished from adjuncts. While a predicate needs its arguments to complete its meaning, the adjuncts that appear with a predicate are optional; they are not necessary to complete the meaning of the predicate. Most theories of syntax and semantics acknowledge arguments and adjuncts, although the terminology varies, and the distinction is generally believed to exist in all languages. Dependency grammars sometimes call arguments actants, following Lucien Tesnière (1959).

In linguistics, a small clause consists of a subject and its predicate, but lacks an overt expression of tense. Small clauses have the semantic subject-predicate characteristics of a clause, and have some, but not all, the properties of a constituent. Structural analyses of small clauses vary according to whether a flat or layered analysis is pursued. The small clause is related to the phenomena of raising-to-object, exceptional case-marking, accusativus cum infinitivo, and object control.

Exceptional case-marking (ECM), in linguistics, is a phenomenon in which the subject of an embedded infinitival verb seems to appear in a superordinate clause and, if it is a pronoun, is unexpectedly marked with object case morphology. The unexpected object case morphology is deemed "exceptional". The term ECM itself was coined in the Government and Binding grammar framework although the phenomenon is closely related to the accusativus cum infinitivo constructions of Latin. ECM-constructions are also studied within the context of raising. The verbs that license ECM are known as raising-to-object verbs. Many languages lack ECM-predicates, and even in English, the number of ECM-verbs is small. The structural analysis of ECM-constructions varies in part according to whether one pursues a relatively flat structure or a more layered one.

In certain theories of linguistics, thematic relations, also known as semantic roles, are the various roles that a noun phrase may play with respect to the action or state described by a governing verb, commonly the sentence's main verb. For example, in the sentence "Susan ate an apple", Susan is the doer of the eating, so she is an agent; an apple is the item that is eaten, so it is a patient.

Syntactic movement is the means by which some theories of syntax address discontinuities. Movement was first postulated by structuralist linguists who expressed it in terms of discontinuous constituents or displacement. Some constituents appear to have been displaced from the position in which they receive important features of interpretation. The concept of movement is controversial and is associated with so-called transformational or derivational theories of syntax. Representational theories, in contrast, reject the notion of movement and often instead address discontinuities with other mechanisms including graph reentrancies, feature passing, and type shifters.

In linguistics, subcategorization denotes the ability/necessity for lexical items to require/allow the presence and types of the syntactic arguments with which they co-occur. The notion of subcategorization is similar to the notion of valency, although the two concepts stem from different traditions in the study of syntax and grammar.