Montague grammar

Last updated

Montague grammar is an approach to natural language semantics, named after American logician Richard Montague. The Montague grammar is based on mathematical logic, especially higher-order predicate logic and lambda calculus, and makes use of the notions of intensional logic, via Kripke models. Montague pioneered this approach in the 1960s and early 1970s.

Contents

Overview

Montague's thesis was that natural languages (like English) and formal languages (like programming languages) can be treated in the same way:

There is in my opinion no important theoretical difference between natural languages and the artificial languages of logicians; indeed, I consider it possible to comprehend the syntax and semantics of both kinds of language within a single natural and mathematically precise theory. On this point I differ from a number of philosophers, but agree, I believe, with Chomsky and his associates. ("Universal Grammar" 1970)

Montague published what soon became known as Montague grammar [1] in three papers:

Illustration

Montague grammar can represent the meanings of quite complex sentences compactly. Below is a grammar presented in Eijck and Unger's textbook. [5]

The types of the syntactic categories in the grammar are as follows, with t denoting a term (a reference to an entity) and f denoting a formula.

categorysymboltype
Sentence S
Verb phrase VP
Noun phrase NP
Common noun CN
Determiner DET
Transitive verb TV

The meaning of a sentence obtained by the rule is obtained by applying the function for NP to the function for VP.

The types of VP and NP might appear unintuitive because of the question as to the meaning of a noun phrase that is not simply a term. This is because meanings of many noun phrases, such as "the man who whistles", are not just terms in predicate logic, but also include a predicate for the activity, like "whistles", which cannot be represented in the term (consisting of constant and function symbols but not of predicates). So we need some term, for example x, and a formula whistles(x) to refer to the man who whistles. The meaning of verb phrases VP can be expressed with that term, for example stating that a particular x satisfies sleeps(x) snores(x) (expressed as a function from x to that formula). Now the function associated with NP takes that kind of function and combines it with the formulas needed to express the meaning of the noun phrase. This particular way of stating NP and VP is not the only possible one.

Key is the meaning of an expression is obtained as a function of its components, either by function application (indicated by boldface parentheses enclosing function and argument) or by constructing a new function from the functions associated with the component. This compositionality makes it possible to assign meanings reliably to arbitrarily complex sentence structures, with auxiliary clauses and many other complications.

The meanings of other categories of expressions are either similarly function applications, or higher-order functions. The following are the rules of the grammar, with the first column indicating a non-terminal symbol, the second column one possible way of producing that non-terminal from other non-terminals and terminals, and the third column indicating the corresponding meaning.

meaning
SNP VP
NPname
NPDET CN
NPDET RCN
DET"some"
DET"a"
DET"every"
DET"no"
VPintransverb
VPTV NP
TVtransverb
RCNCN "that" VP
RCNCN "that" NP TV
CNpredicate

Here are example expressions and their associated meaning, according to the above grammar, showing that the meaning of a given sentence is formed from its constituent expressions, either by forming a new higher-order function, or by applying a higher-order function for one expression to the meaning of another.

expressionmeaning
a
man
a man
sleeps
a man sleeps
man that dreams
a man that dreams
a man that dreams sleeps

The following are other examples of sentences translated into the predicate logic by the grammar.

sentencetranslation to logic
Jill sees Jack
every woman sees a man
every woman sees a man that sleeps
a woman that eats sees a man that sleeps

In David Foster Wallace's novel Infinite Jest , the protagonist Hal Incandenza has written an essay entitled Montague Grammar and the Semantics of Physical Modality. Montague grammar is also referenced explicitly and implicitly several times throughout the book.

See also

Related Research Articles

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). There are numerous approaches to syntax that differ in their central assumptions and goals.

Semantics is the study of reference, meaning, or truth. The term can be used to refer to subfields of several distinct disciplines, including philosophy, linguistics and computer science.

<span class="mw-page-title-main">Parse tree</span>

A parse tree or parsing tree or derivation tree or concrete syntax tree is an ordered, rooted tree that represents the syntactic structure of a string according to some context-free grammar. The term parse tree itself is used primarily in computational linguistics; in theoretical syntax, the term syntax tree is more common.

Lexical functional grammar (LFG) is a constraint-based grammar framework in theoretical linguistics. It posits two separate levels of syntactic structure, a phrase structure grammar representation of word order and constituency, and a representation of grammatical functions such as subject and object, similar to dependency grammar. The development of the theory was initiated by Joan Bresnan and Ronald Kaplan in the 1970s, in reaction to the theory of transformational grammar which was current in the late 1970s. It mainly focuses on syntax, including its relation with morphology and semantics. There has been little LFG work on phonology.

Metalogic is the study of the metatheory of logic. Whereas logic studies how logical systems can be used to construct valid and sound arguments, metalogic studies the properties of logical systems. Logic concerns the truths that may be derived using a logical system; metalogic concerns the truths that may be derived about the languages and systems that are used to express truths.

Categorial grammar is a family of formalisms in natural language syntax that share the central assumption that syntactic constituents combine as functions and arguments. Categorial grammar posits a close relationship between the syntax and semantic composition, since it typically treats syntactic categories as corresponding to semantic types. Categorial grammars were developed in the 1930s by Kazimierz Ajdukiewicz and in the 1950s by Yehoshua Bar-Hillel and Joachim Lambek. It saw a surge of interest in the 1970s following the work of Richard Montague, whose Montague grammar assumed a similar view of syntax. It continues to be a major paradigm, particularly within formal semantics.

In generative grammar, a theta role or θ-role is the formal device for representing syntactic argument structure—the number and type of noun phrases—required syntactically by a particular verb. For example, the verb put requires three arguments.

The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue. Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining trait of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.

The term predicate is used in one of two ways in linguistics and its subfields. The first defines a predicate as everything in a standard declarative sentence except the subject, and the other views it as just the main content verb or associated predicative expression of a clause. Thus, by the first definition the predicate of the sentence Frank likes cake is likes cake. By the second definition, the predicate of the same sentence is just the content verb likes, whereby Frank and cake are the arguments of this predicate. Differences between these two definitions can lead to confusion.

In formal semantics and philosophy of language, a meaning postulate is a way of stipulating a relationship between the meanings of two or more words. They were introduced by Rudolf Carnap as a way of approaching the analytic/synthetic distinction. Subsequently, Richard Montague made heavy use of meaning postulates in the development of Montague grammar, and they have features prominently in formal semantics following in Montague's footsteps.

A definite clause grammar (DCG) is a way of expressing grammar, either for natural or formal languages, in a logic programming language such as Prolog. It is closely related to the concept of attribute grammars / affix grammars from which Prolog was originally developed. DCGs are usually associated with Prolog, but similar languages such as Mercury also include DCGs. They are called definite clause grammars because they represent a grammar as a set of definite clauses in first-order logic.

In formal linguistics, discourse representation theory (DRT) is a framework for exploring meaning under a formal semantics approach. One of the main differences between DRT-style approaches and traditional Montagovian approaches is that DRT includes a level of abstract mental representations within its formalism, which gives it an intrinsic ability to handle meaning across sentence boundaries. DRT was created by Hans Kamp in 1981. A very similar theory was developed independently by Irene Heim in 1982, under the name of File Change Semantics (FCS). Discourse representation theories have been used to implement semantic parsers and natural language understanding systems.

In generative grammar and related approaches, the logical form (LF) of a linguistic expression is the variant of its syntactic structure which undergoes semantic interpretation. It is distinguished from phonetic form, the structure which corresponds to a sentence's pronunciation. These separate representations are postulated in order to explain the ways in which an expression's meaning can be partially independent of its pronunciation, e.g. scope ambiguities.

In linguistics, an argument is an expression that helps complete the meaning of a predicate, the latter referring in this context to a main verb and its auxiliaries. In this regard, the complement is a closely related concept. Most predicates take one, two, or three arguments. A predicate and its arguments form a predicate-argument structure. The discussion of predicates and arguments is associated most with (content) verbs and noun phrases (NPs), although other syntactic categories can also be construed as predicates and as arguments. Arguments must be distinguished from adjuncts. While a predicate needs its arguments to complete its meaning, the adjuncts that appear with a predicate are optional; they are not necessary to complete the meaning of the predicate. Most theories of syntax and semantics acknowledge arguments and adjuncts, although the terminology varies, and the distinction is generally believed to exist in all languages. Dependency grammars sometimes call arguments actants, following Lucien Tesnière (1959).

In linguistics, inverse copular constructions, named after Moro (1997), are a type of inversion in English where canonical SCP word order is reversed in a sense, so that one appears to have the order PCS instead. The verb in these constructions is always the copula be. Inverse copular constructions are intriguing because they render the distinction between subject and predicative expression difficult to maintain. The confusion has led to focused study of these constructions, and their impact on the theory of grammar may be great since they appear to challenge the initial binary division of the sentence (S) into a subject noun phrase (NP) and a predicate verb phrase (VP), this division being at the core of all phrase structure grammars.

Donkey sentences are sentences that contain a pronoun with clear meaning but whose syntactical role in the sentence poses challenges to grammarians. Such sentences defy straightforward attempts to generate their formal language equivalents. The difficulty is with understanding how English speakers parse such sentences.

<span class="mw-page-title-main">Symbol (formal)</span>

A logical symbol is a fundamental concept in logic, tokens of which may be marks or a configuration of marks which form a particular pattern. Although the term "symbol" in common use refers at some times to the idea being symbolized, and at other times to the marks on a piece of paper or chalkboard which are being used to express that idea; in the formal languages studied in mathematics and logic, the term "symbol" refers to the idea, and the marks are considered to be a token instance of the symbol. In logic, symbols build literal utility to illustrate ideas.

Discontinuous-constituent Phrase Structure Grammar (DCPSG) is a formalism for describing discontinuous phrase structures in natural language, such as verb phrases in VSO languages. The formalism was introduced in the slightly more constrained form of Discontinuous-constituent Phrase Structure Grammar with Subscripts and Deletes (DCPSGsd) in Harman (1963). DCPSGs describe a superset of the context-free languages, by means of rewrite rules that permit a limited amount of wrapping, similar to that found in Head grammar.

Formal semantics is the study of grammatical meaning in natural languages using formal tools from logic and theoretical computer science. It is an interdisciplinary field, sometimes regarded as a subfield of both linguistics and philosophy of language. It provides accounts of what linguistic expressions mean and how their meanings are composed from the meanings of their parts. The enterprise of formal semantics can be thought of as that of reverse-engineering the semantic components of natural languages' grammars.

The Temperature Paradox or Partee's Paradox is a problem in the formal semantics of natural language and philosophical logic. Formulated by Barbara Partee, it consists of the following argument, which would be wrongly predicted as valid by many formalizations.

  1. The temperature is rising.
  2. The temperature is ninety.
  3. Therefore, ninety is rising.

References

  1. The linguist Barbara Partee credibly claims to have invented the term in 1971 “for the system spelled out in Montague's“ UG, EFL and “especially in PTQ”. See her essay "Reflections of a Formal Semanticist as of Feb 2005", p. 14, footnote 36.
  2. "Universal grammar". Theoria 36 (1970), 373–398. (reprinted in Thomason, 1974)
  3. "English as a Formal Language". In: Bruno Visentini (ed.): Linguaggi nella società e nella tecnica. Mailand 1970, 189–223. (reprinted in Thomason, 1974)
  4. "The Proper Treatment of Quantification in Ordinary English". In: Jaakko Hintikka, Julius Moravcsik, Patrick Suppes (eds.): Approaches to Natural Language. Dordrecht 1973, 221–242. (reprinted in Thomason, 1974)
  5. J. van Eijck and C. Unger. Computational Semantics with Functional Programming. Cambridge University Press, 2010.

Further reading