Formal linguistics

Last updated

Formal linguistics is the branch of linguistics which uses applied mathematical methods for the analysis of natural languages. Such methods include formal languages, formal grammars and first-order logical expressions. Formal linguistics also forms the basis of computational linguistics. Since the 1980s, the term is often used to refer to Chomskyan linguistics. [1]

Contents

Approaches

Semiotic

Methods of formal linguistics were introduced by semioticians such as Charles Sanders Peirce and Louis Hjelmslev. Building on the work of David Hilbert and Rudolf Carnap, Hjelmslev proposed the use of formal grammars to analyse, generate and explain language in his 1943 book Prolegomena to a Theory of Language. [2] [3] In this view, language is regarded as arising from a mathematical relationship between meaning and form.

The formal description of language was further developed by linguists including J. R. Firth and Simon Dik, giving rise to modern grammatical frameworks such as systemic functional linguistics and functional discourse grammar. Computational methods have been developed by the framework functional generative description among others.

Dependency grammar, created by French structuralist Lucien Tesnière, [4] has been used widely in natural language processing.

Psychological

Analytical models based on semantics and discourse pragmatics were rejected by the Bloomfieldian school of linguistics [5] whose derivatives place the object into the verb phrase, following from Wilhelm Wundt's Völkerpsychologie. Formalisms based on this convention were constructed in the 1950s by Zellig Harris and Charles Hockett. These gave rise to modern generative grammar. [3] It has been suggested that dependency relations are caused by a random mutation in the human genome. [6]

Generative models of formal linguistics, such as head-driven phrase structure grammar, have also been used in natural language processing.

See also

Related Research Articles

In linguistics, the grammar of a natural language is its set of structural rules on speakers' or writers' usage and creation of clauses, phrases, and words. The term can also refer to the study of such rules, a subject that includes phonology, morphology, and syntax, together with phonetics, semantics, and pragmatics. There are two different ways to study grammar right now: traditional grammar and theoretical grammar.

The following outline is provided as an overview and topical guide to linguistics:

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). There are numerous approaches to syntax that differ in their central assumptions and goals.

Semantics is the study of reference, meaning, or truth. The term can be used to refer to subfields of several distinct disciplines, including philosophy, linguistics and computer science.

In linguistics, transformational grammar (TG) or transformational-generative grammar (TGG) is part of the theory of generative grammar, especially of natural languages. It considers grammar to be a system of rules that generate exactly those combinations of words that form grammatical sentences in a given language and involves the use of defined operations to produce new sentences from existing ones. The method is commonly associated with American linguist Noam Chomsky.

<span class="mw-page-title-main">Generative grammar</span> Theory in linguistics

Generative grammar, or generativism, is a linguistic theory that regards linguistics as the study of a hypothesised innate grammatical structure. It is a biological or biologistic modification of earlier structuralist theories of linguistics, deriving ultimately from glossematics. Generative grammar considers grammar as a system of rules that generates exactly those combinations of words that form grammatical sentences in a given language. It is a system of explicit rules that may apply repeatedly to generate an indefinite number of sentences which can be as long as one wants them to be. The difference from structural and functional models is that the object is base-generated within the verb phrase in generative grammar. This purportedly cognitive structure is thought of as being a part of a universal grammar, a syntactic structure which is caused by a genetic mutation in humans.

Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.

The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue. Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining trait of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.

<i>Syntactic Structures</i> Book by Noam Chomsky

Syntactic Structures is an important work in linguistics by American linguist Noam Chomsky, originally published in 1957. A short monograph of about a hundred pages, it is recognized as one of the most significant and influential linguistic studies of the 20th century. It contains the now-famous sentence "Colorless green ideas sleep furiously", which Chomsky offered as an example of a grammatically correct sentence that has no discernible meaning, thus arguing for the independence of syntax from semantics.

Generative semantics was a research program in theoretical linguistics which held that syntactic structures are computed on the basis of meanings rather than the other way around. Generative semantics developed out of transformational generative grammar in the mid-1960s, but stood in opposition to it. The period in which the two research programs coexisted was marked by intense and often personal clashes now known as the linguistics wars. Its proponents included Haj Ross, Paul Postal, James McCawley, and George Lakoff, who dubbed themselves "The Four Horsemen of the Apocalypse".

Glossematics is a structuralist linguistic theory proposed by Louis Hjelmslev and Hans Jørgen Uldall although the two ultimately went separate ways each with their own approach. Hjelmslev’s theory, most notably, is an early mathematical methodology for the analysis of language which was subsequently incorporated into the analytical foundation of current models of functional—structural grammar such as Danish Functional Grammar, Functional Discourse Grammar and Systemic Functional Linguistics. Hjelmslev’s theory likewise remains fundamental for modern semiotics.

<span class="mw-page-title-main">Lucien Tesnière</span> French linguist

Lucien Tesnière was a prominent and influential French linguist. He was born in Mont-Saint-Aignan on May 13, 1893. As a senior lecturer at the University of Strasbourg (1924) and later professor at the University of Montpellier (1937), he published many papers and books on Slavic languages. However, his importance in the history of linguistics is based mainly on his development of an approach to the syntax of natural languages that would become known as dependency grammar. He presented his theory in his book Éléments de syntaxe structurale, published posthumously in 1959. In the book he proposes a sophisticated formalization of syntactic structures, supported by many examples from a diversity of languages. Tesnière died in Montpellier on December 6, 1954.

In linguistics, an argument is an expression that helps complete the meaning of a predicate, the latter referring in this context to a main verb and its auxiliaries. In this regard, the complement is a closely related concept. Most predicates take one, two, or three arguments. A predicate and its arguments form a predicate-argument structure. The discussion of predicates and arguments is associated most with (content) verbs and noun phrases (NPs), although other syntactic categories can also be construed as predicates and as arguments. Arguments must be distinguished from adjuncts. While a predicate needs its arguments to complete its meaning, the adjuncts that appear with a predicate are optional; they are not necessary to complete the meaning of the predicate. Most theories of syntax and semantics acknowledge arguments and adjuncts, although the terminology varies, and the distinction is generally believed to exist in all languages. Dependency grammars sometimes call arguments actants, following Lucien Tesnière (1959).

Langueandparole is a theoretical linguistic dichotomy distinguished by Ferdinand de Saussure in his Course in General Linguistics.

Donkey sentences are sentences that contain a pronoun with clear meaning but whose syntactical role in the sentence poses challenges to grammarians. Such sentences defy straightforward attempts to generate their formal language equivalents. The difficulty is with understanding how English speakers parse such sentences.

Structural linguistics, or structuralism, in linguistics, denotes schools or theories in which language is conceived as a self-contained, self-regulating semiotic system whose elements are defined by their relationship to other elements within the system. It is derived from the work of Swiss linguist Ferdinand de Saussure and is part of the overall approach of structuralism. Saussure's Course in General Linguistics, published posthumously in 1916, stressed examining language as a dynamic system of interconnected units. Saussure is also known for introducing several basic dimensions of semiotic analysis that are still important today. Two of these are his key methods of syntagmatic and paradigmatic analysis, which define units syntactically and lexically, respectively, according to their contrast with the other units in the system.

The Copenhagen School is a group of scholars dedicated to the study of linguistics, centered around Louis Hjelmslev (1899–1965) and the Linguistic Circle of Copenhagen, founded by him and Viggo Brøndal (1887–1942). In the mid twentieth century the Copenhagen school was one of the most important centres of linguistic structuralism together with the Geneva School and the Prague School. In the late 20th and early 21st century the Copenhagen school has turned from a purely structural approach to linguistics to a functionalist one, Danish functional linguistics, which nonetheless incorporates many insights from the founders of the Linguistic Circle of Copenhagen.

Linguistics is the scientific study of language. The modern-day scientific study of linguistics takes all aspects of language into account — i.e., the cognitive, the social, the cultural, the psychological, the environmental, the biological, the literary, the grammatical, the paleographical, and the structural. Linguistics is based on the theoretical as well as descriptive study of language, and is also interlinked with the applied fields of language studies and language learning, which entails the study of specific languages. Before the 20th century, linguistics evolved in an informal manner that did not employ scientific methods.

<span class="mw-page-title-main">Formalism (linguistics)</span> Concept in linguistics

In linguistics, the term formalism is used in a variety of meanings which relate to formal linguistics in different ways. In common usage, it is merely synonymous with a grammatical model or a syntactic model: a method for analyzing sentence structures. Such formalisms include different methodologies of generative grammar which are especially designed to produce grammatically correct strings of words; or the likes of Functional Discourse Grammar which builds on predicate logic.

Theory of language is a topic from philosophy of language and theoretical linguistics. It has the goal of answering the questions "What is language?"; "Why do languages have the properties they have?"; or "What is the origin of language?". In addition to these fundamental questions, the theory of language also seeks to understand how language is acquired and used by individuals and communities. This involves investigating the cognitive and neural processes involved in language processing and production, as well as the social and cultural factors that shape linguistic behavior.

References

  1. Haspelmath, Martin. "How formal linguistics appeared and disappeared from the scene".
  2. Hjelmslev, Louis (1969) [First published 1943]. Prolegomena to a Theory of Language. University of Wisconsin Press. ISBN   0299024709.
  3. 1 2 Seuren, Pieter A. M. (1998). Western linguistics: An historical introduction. Wiley-Blackwell. pp. 160–167. ISBN   0-631-20891-7.
  4. Tesnière, Lucien (1959). Éléments de syntaxe structurale. Klincksieck.
  5. Garvin, Paul L. (1954). "Prolegomena to a Theory of Language by Louis Hjelmslev; Francis J. Whitfield". Language. 30 (1): 69–96. doi:10.2307/410221. JSTOR   410221.
  6. Berwick, Robert C.; Chomsky, Noam (2015). Why Only Us: Language and Evolution. MIT Press. ISBN   9780262034241.