Theta role

Last updated

Theta roles are the names of the participant roles associated with a predicate: the predicate may be a verb, an adjective, a preposition, or a noun. If an object is in motion or in a steady state as the speakers perceives the state, or it is the topic of discussion, it is called a theme. [1] The[ clarification needed ] participant is usually said to be an argument of the predicate. In generative grammar, a theta role or θ-role is the formal device for representing syntactic argument structure—the number and type of noun phrases—required syntactically by a particular verb. For example, the verb put requires three arguments (i.e., it is trivalent)[ citation needed ].

Contents

The formal mechanism for implementing a verb's argument structure is codified as theta roles. The verb put is said to "assign" three theta roles. This is coded in a theta grid associated with the lexical entry for the verb. The correspondence between the theta grid and the actual sentence is accomplished by means of a bijective filter on the grammar known as the theta criterion. Early conceptions of theta roles include Fillmore (1968) (Fillmore called theta roles "cases") and Gruber (1965).

Theta roles are prominent in government and binding theory and the standard theory of transformational grammar.

Thematic relations

The term "theta role" is often used interchangeably with the term thematic relations (particularly in mainstream generative grammar—for an exception see Carnie 2006). The reason for this is simple: theta roles typically reference thematic relations. In particular, theta roles are often referred to by the most prominent thematic relation in them. For example, a common theta role is the primary or external argument. Typically, although not always, this theta role maps to a noun phrase which bears an agent thematic relation. As such, the theta role is called the "agent" theta role. This often leads to confusion between the two notions. The two concepts, however, can be distinguished in a number of ways.

is a fact about theta roles (the number and types of the arguments). The actual semantic type of the argument is described by the thematic relation.

One common way of thinking about theta roles is that they are bundles of thematic relations associated with a particular argument position ( Carnie 2006 ).

Theta grids and the theta criterion

Theta roles are stored in a verb's theta grid. Grids typically come in two forms. The simplest and easiest to type is written as an ordered list between angle brackets. The argument associated with the external argument position (which typically ends up being the subject in active sentences) is written first and underlined. The theta roles are named by the most prominent thematic relation that they contain. In this notation, the theta grid for a verb such as give is <agent, theme, goal>.

The other notation (see for example the textbook examples in Haegeman 1994 and Carnie 2006) separates the theta roles into boxes, in which each column represents a theta role. The top row represents the names of the thematic relations contained in the theta role. In some work (e.g., Carnie 2006), this box also contains information about the category associated with the theta role. This mingles theta-theory with the notion of subcategorization. The bottom row gives a series of indexes which are associated with subscripted markers in the sentence itself which indicate that the NPs they are attached to have been assigned the theta role in question.

The theta grid for give
Agent
source
DP
theme
DP
goal
PP
ijk

[ clarification needed ]

When applied to the sentence [S[NP Susan]i gave [NP the food]j [PPto Biff]k] the indices mark that Susan is assigned the external theta role of agent/source, the food is assigned the theme role, and to Biff is assigned the goal role.

The theta criterion (or θ-criterion) is the formal device in Government and Binding Theory for enforcing the one to one match between arguments and theta roles. This acts as a filter on the D-structure of the sentence. If an[ clarification needed ] argument fails to have the correct match between the number of arguments (typically NPs, PPs, or embedded clauses) and the number of theta roles, the sentence will be ungrammatical or unparseable. Chomsky's formulation ( Chomsky 1981 , p. 36) is:

The theta criterion Each argument bears one and only one θ-role, and each θ-role is assigned to one and only one argument.

Although it is often not explicitly stated, adjuncts are excluded from the theta criterion.

Thematic hierarchies

Drawing on observations based in typological cross-linguistic comparisons of languages ( Fillmore 1968 ), linguists in the relational grammar (RG) tradition (e.g. Perlmutter & Postal 1984) observed that particular thematic relations and theta roles map on to particular positions in the sentence. For example, in unmarked situations agents map to subject positions, themes onto object position, and goals onto indirect objects. In RG, this is encoded in the Universal Alignment Hypothesis (or UAH), where the thematic relations are mapped directly into argument position based on the following hierarchy: Agent < Theme < Experiencer < Others. Mark Baker adopted this idea into GB theory in the form of the Uniformity of Theta Assignment Hypothesis (or UTAH) ( Baker 1988 ). UTAH explains how identical thematic relationships between items are shown by identical structural relationships. A different approach to the correspondence is given in ( Hale & Keyser 1993 ) and ( Hale & Keyser 2001 ), where there are no such things as underlying theta roles or even thematic relations. Instead, the interpretive component of the grammar identifies the semantic role of an argument based on its position in the tree.

Argument structure in other formal approaches

Lexical-functional grammar (LFG)

Lexical-functional grammar (LFG) ( Falk 2001 ) and ( Bresnan 2001 ) is perhaps the most similar to Chomskyan approaches in implementing theta-roles. However, LFG uses three distinct layers of structure for representing the relations or functions of arguments: θ-structure, a-structure (argument structure) and f-structure (functional structure) which expresses[ clarification needed ] grammatical relations. These three layers are linked together using a set of intricate linking principles. Thematic relations in the θ-structure are mapped onto a set of positions in the a-structure which are tied to features [+o] (roughly "object") and [±r] (roughly "restricted" meaning it is marked explicitly by a preposition or a case marking). Themes map to [-r], second themes map to [+o] and non-themes map to [-o]. These features then determine how the arguments are mapped to specific grammatical functions in the sentence. The first [-o] argument is mapped to the SUBJ (subject) relation. If there is no [-o] argument then the first [-r] argument is mapped to the SUBJ relation. If neither of these apply, then you add the plus value ([+r] or [+o]) to the feature structure and apply the following mappings: [-o,-r]: SUBJ, [+o, -r]: Object (OBJ), [-o,+r]: prepositional marked oblique (OBLθ), [+o, +r]: prepositionally marked object (OBJθ). These mappings are further constrained by the following constraints:

Function argument biuniqueness: Each a-structure role corresponds to a unique f-structure function, and each f-structure function corresponds to a unique a-structure role

The Subject Condition: Every verb must have a SUBJ

F-structures are further constrained by the following two constraints which do much of the same labor as the θ-criterion:

Coherence requires that every participant in the f-structure of a sentence must be mentioned in a-structure (or in a constituting equation) of a predicate in its clause.

Completeness: An f-structure for a sentence must contain values for all the grammatical functions mentioned in a-structure.

Head-driven phrase structure grammar (HPSG)

Head-driven phrase structure grammar (HPSG) (for a textbook introduction, see Sag, Wasow & Bender 2005) does not use theta roles per se, but divides their property into two distinct feature structures. The number and category are indicated by a feature called ARG-STR. This feature is an ordered list of categories that must cooccur with a particular verb or predicate. For example, the ARG-STR list of the verb give is <NP, NP, PP>. The semantic part of theta roles (i.e. the thematic relations) are treated in a special set of semantic restriction (RESTR) features. These typically express the semantic properties more directly than thematic relations. For example, the semantic relations associated with the arguments of the verb give are not agent, theme and goal, but giver, given, givee.

Approaches that eschew theta roles

Many approaches to grammar including construction grammar and the Simpler Syntax model ( Culicover & Jackendoff 2005 ) (see also Jackendoff's earlier work on argument structure and semantics, including Jackendoff 1983 and Jackendoff 1990) claim that theta roles (and thematic relations) are neither a good way to represent the syntactic argument structure of predicates nor of the semantic properties that they reveal. They argue for more complex and articulated semantic structures (often called Lexical-conceptual structures) which map onto the syntactic structure.

Similarly, most typological approaches to grammar, functionalist theories (such as functional grammar and Role and Reference Grammar ( Van Valin & LaPolla 1997 ), and dependency grammar do not use theta roles, but they may make reference to thematic relations and grammatical relations or their notational equivalents. These are usually related to one another directly using principles of mapping.

See also

Related Research Articles

<span class="mw-page-title-main">Syntax</span> System responsible for combining morphemes into complex structures

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). There are numerous approaches to syntax that differ in their central assumptions and goals.

A syntactic category is a syntactic unit that theories of syntax assume. Word classes, largely corresponding to traditional parts of speech, are syntactic categories. In phrase structure grammars, the phrasal categories are also syntactic categories. Dependency grammars, however, do not acknowledge phrasal categories.

An adjective phrase is a phrase whose head is an adjective. Almost any grammar or syntax textbook or dictionary of linguistics terminology defines the adjective phrase in a similar way, e.g. Kesner Bland (1996:499), Crystal (1996:9), Greenbaum (1996:288ff.), Haegeman and Guéron (1999:70f.), Brinton (2000:172f.), Jurafsky and Martin (2000:362). The adjective can initiate the phrase, conclude the phrase, or appear in a medial position. The dependents of the head adjective—i.e. the other words and phrases inside the adjective phrase—are typically adverb or prepositional phrases, but they can also be clauses. Adjectives and adjective phrases function in two basic ways, attributively or predicatively. An attributive adjective (phrase) precedes the noun of a noun phrase. A predicative adjective (phrase) follows a linking verb and serves to describe the preceding subject, e.g. The man is very happy.

Case roles, according to the work by Fillmore (1967), are the semantic roles of noun phrases in relation to the syntactic structures that contain these noun phrases. The term case role is most widely used for purely semantic relations, including theta roles and thematic roles, that can be independent of the morpho-syntax. The concept of case roles is related to the larger notion of Case which is defined as a system of marking dependent nouns for the type of semantic or syntactic relationship they bear to their heads. Case traditionally refers to inflectional marking.

Lexical semantics, as a subfield of linguistic semantics, is the study of word meanings. It includes the study of how words structure their meaning, how they act in grammar and compositionality, and the relationships between the distinct senses and uses of a word.

Lexical functional grammar (LFG) is a constraint-based grammar framework in theoretical linguistics. It posits two separate levels of syntactic structure, a phrase structure grammar representation of word order and constituency, and a representation of grammatical functions such as subject and object, similar to dependency grammar. The development of the theory was initiated by Joan Bresnan and Ronald Kaplan in the 1970s, in reaction to the theory of transformational grammar which was current in the late 1970s. It mainly focuses on syntax, including its relation with morphology and semantics. There has been little LFG work on phonology.

Head-driven phrase structure grammar (HPSG) is a highly lexicalized, constraint-based grammar developed by Carl Pollard and Ivan Sag. It is a type of phrase structure grammar, as opposed to a dependency grammar, and it is the immediate successor to generalized phrase structure grammar. HPSG draws from other fields such as computer science and uses Ferdinand de Saussure's notion of the sign. It uses a uniform formalism and is organized in a modular way which makes it attractive for natural language processing.

<span class="mw-page-title-main">Generative grammar</span> Theory in linguistics

Generative grammar, or generativism, is a linguistic theory that regards linguistics as the study of a hypothesised innate grammatical structure. It is a biological or biologistic modification of earlier structuralist theories of linguistics, deriving from logical syntax and glossematics. Generative grammar considers grammar as a system of rules that generates exactly those combinations of words that form grammatical sentences in a given language. It is a system of explicit rules that may apply repeatedly to generate an indefinite number of sentences which can be as long as one wants them to be. The difference from structural and functional models is that the object is base-generated within the verb phrase in generative grammar. This purportedly cognitive structure is thought of as being a part of a universal grammar, a syntactic structure which is caused by a genetic mutation in humans.

In linguistics, valency or valence is the number and type of arguments controlled by a predicate, content verbs being typical predicates. Valency is related, though not identical, to subcategorization and transitivity, which count only object arguments – valency counts all arguments, including the subject. The linguistic meaning of valency derives from the definition of valency in chemistry. Like valency found in chemistry, there is the binding of specific elements. In the grammatical theory of valency, the verbs organize sentences by binding the specific elements. Examples of elements that would be bound would be the complement and the actant. Although the term originates from valence in chemistry, linguistic valency has a close analogy in mathematics under the term arity.

In generative grammar and related frameworks, a node in a parse tree c-commands its sister node and all of its sister's descendants. In these frameworks, c-command plays a central role in defining and constraining operations such as syntactic movement, binding, and scope. Tanya Reinhart introduced c-command in 1976 as a key component of her theory of anaphora. The term is short for "constituent command".

The term predicate is used in one of two ways in linguistics and its subfields. The first defines a predicate as everything in a standard declarative sentence except the subject, and the other views it as just the main content verb or associated predicative expression of a clause. Thus, by the first definition the predicate of the sentence Frank likes cake is likes cake. By the second definition, the predicate of the same sentence is just the content verb likes, whereby Frank and cake are the arguments of this predicate. Differences between these two definitions can lead to confusion.

<span class="mw-page-title-main">Grammatical relation</span>

In linguistics, grammatical relations are functional relationships between constituents in a clause. The standard examples of grammatical functions from traditional grammar are subject, direct object, and indirect object. In recent times, the syntactic functions, typified by the traditional categories of subject and object, have assumed an important role in linguistic theorizing, within a variety of approaches ranging from generative grammar to functional and cognitive theories. Many modern theories of grammar are likely to acknowledge numerous further types of grammatical relations. The role of grammatical relations in theories of grammar is greatest in dependency grammars, which tend to posit dozens of distinct grammatical relations. Every head-dependent dependency bears a grammatical function.

The theta-criterion is a constraint on x-bar theory that was first proposed by Noam Chomsky (1981) as a rule within the system of principles of the government and binding theory, called theta-theory (θ-theory). As theta-theory is concerned with the distribution and assignment of theta-roles, the theta-criterion describes the specific match between arguments and theta-roles (θ-roles) in logical form (LF):

In generative grammar and related approaches, the logical form (LF) of a linguistic expression is the variant of its syntactic structure which undergoes semantic interpretation. It is distinguished from phonetic form, the structure which corresponds to a sentence's pronunciation. These separate representations are postulated in order to explain the ways in which an expression's meaning can be partially independent of its pronunciation, e.g. scope ambiguities.

In linguistics, an argument is an expression that helps complete the meaning of a predicate, the latter referring in this context to a main verb and its auxiliaries. In this regard, the complement is a closely related concept. Most predicates take one, two, or three arguments. A predicate and its arguments form a predicate-argument structure. The discussion of predicates and arguments is associated most with (content) verbs and noun phrases (NPs), although other syntactic categories can also be construed as predicates and as arguments. Arguments must be distinguished from adjuncts. While a predicate needs its arguments to complete its meaning, the adjuncts that appear with a predicate are optional; they are not necessary to complete the meaning of the predicate. Most theories of syntax and semantics acknowledge arguments and adjuncts, although the terminology varies, and the distinction is generally believed to exist in all languages. Dependency grammars sometimes call arguments actants, following Lucien Tesnière (1959).

In linguistics, volition is a concept that distinguishes whether the subject, or agent of a particular sentence intended an action or not. Simply, it is the intentional or unintentional nature of an action. Volition concerns the idea of control and for the purposes outside of psychology and cognitive science, is considered the same as intention in linguistics. Volition can then be expressed in a given language using a variety of possible methods. These sentence forms usually indicate that a given action has been done intentionally, or willingly. There are various ways of marking volition cross-linguistically. When using verbs of volition in English, like "want" or "prefer", these verbs are not expressly marked. Other languages handle this with affixes, while others have complex structural consequences of volitional or non-volitional encoding.

In certain theories of linguistics, thematic relations, also known as semantic roles, are the various roles that a noun phrase may play with respect to the action or state described by a governing verb, commonly the sentence's main verb. For example, in the sentence "Susan ate an apple", Susan is the doer of the eating, so she is an agent; an apple is the item that is eaten, so it is a patient.

In linguistics, subcategorization denotes the ability/necessity for lexical items to require/allow the presence and types of the syntactic arguments with which they co-occur. For example, the word "walk" as in "X walks home" requires the noun-phrase X to be animate.

In linguistics, selection denotes the ability of predicates to determine the semantic content of their arguments. Predicates select their arguments, which means they limit the semantic content of their arguments. One sometimes draws a distinction between types of selection; one acknowledges both s(emantic)-selection and c(ategory)-selection. Selection in general stands in contrast to subcategorization: predicates both select and subcategorize for their complement arguments, whereas they only select their subject arguments. Selection is a semantic concept, whereas subcategorization is a syntactic one. Selection is closely related to valency, a term used in other grammars than the Chomskian generative grammar, for a similar phenomenon.

In linguistics, the syntax–semantics interface is the interaction between syntax and semantics. Its study encompasses phenomena that pertain to both syntax and semantics, with the goal of explaining correlations between form and meaning. Specific topics include scope, binding, and lexical semantic properties such as verbal aspect and nominal individuation, semantic macroroles, and unaccusativity.

References

  1. "Theta Roles". www.sfu.ca. Retrieved 2023-05-20.

Sources