Link grammar

Last updated

Link grammar (LG) is a theory of syntax by Davy Temperley and Daniel Sleator which builds relations between pairs of words, rather than constructing constituents in a phrase structure hierarchy. Link grammar is similar to dependency grammar, but dependency grammar includes a head-dependent relationship, whereas link grammar makes the head-dependent relationship optional (links need not indicate direction). [1] Colored Multiplanar Link Grammar (CMLG) is an extension of LG allowing crossing relations between pairs of words. [2] The relationship between words is indicated with link types, thus making the Link grammar closely related to certain categorial grammars.

Contents

For example, in a subject–verb–object language like English, the verb would look left to form a subject link, and right to form an object link. Nouns would look right to complete the subject link, or left to complete the object link.

In a subject–object–verb language like Persian, the verb would look left to form an object link, and a more distant left to form a subject link. Nouns would look to the right for both subject and object links.

Overview

Link grammar connects the words in a sentence with links, similar in form to a catena. Unlike the catena or a traditional dependency grammar, the marking of the head-dependent relationship is optional for most languages, becoming mandatory only in free-word-order languages (such as Turkish, [3] [ better source needed ] Finnish, Hungarian). That is, in English, the subject-verb relationship is "obvious", in that the subject is almost always to the left of the verb, and thus no specific indication of dependency needs to be made. In the case of subject-verb inversion, a distinct link type is employed. For free word-order languages, this can no longer hold, and a link between the subject and verb must contain an explicit directional arrow to indicate which of the two words is which.

Link grammar also differs from traditional dependency grammars by allowing cyclic relations between words. Thus, for example, there can be links indicating both the head verb of a sentence, the head subject of the sentence, as well as a link between the subject and the verb. These three links thus form a cycle (a triangle, in this case). Cycles are useful in constraining what might otherwise be ambiguous parses; cycles help "tighten up" the set of allowable parses of a sentence.

For example, in the parse

    +---->WV--->+            +--Wd--+-Ss-+--Pa--+     |      |    |      | LEFT-WALL he  runs   fast

the LEFT-WALL indicates the start of the sentence, or the root node. The directional WV link (with arrows) points at the head verb of the sentence; it is the Wall-Verb link. [4] The Wd link (drawn here without arrows) indicates the head noun (the subject) of the sentence. The link type Wd indicates both that it connects to the wall (W) and that the sentence is a declarative sentence (the lower-case "d" subtype). [5] The Ss link indicates the subject-verb relationship; the lower-case "s" indicating that the subject is singular. [6] Note that the WV, Wd and Ss links for a cycle. The Pa link connects the verb to a complement; the lower-case "a" indicating that it is a predicative adjective in this case. [7]

Parsing algorithm

Parsing is performed in analogy to assembling a jigsaw puzzle (representing the parsed sentence) from puzzle pieces (representing individual words). [8] [9] A language is represented by means of a dictionary or lexis, which consists of words and the set of allowed "jigsaw puzzle shapes" that each word can have. The shape is indicated by a "connector", which is a link-type, and a direction indicator + or - indicating right or left. Thus for example, a transitive verb may have the connectors S- & O+ indicating that the verb may form a Subject ("S") connection to its left ("-") and an object connection ("O") to its right ("+"). Similarly, a common noun may have the connectors D- & S+ indicating that it may connect to a determiner on the left ("D-") and act as a subject, when connecting to a verb on the right ("S+"). The act of parsing is then to identify that the S+ connector can attach to the S- connector, forming an "S" link between the two words. Parsing completes when all connectors have been connected.

A given word may have dozens or even hundreds of allowed puzzle-shapes (termed "disjuncts"): for example, many verbs may be optionally transitive, thus making the O+ connector optional; such verbs might also take adverbial modifiers (E connectors) which are inherently optional. More complex verbs may have additional connectors for indirect objects, or for particles or prepositions. Thus, a part of parsing also involves picking one single unique disjunct for a word; the final parse must satisfy (connect) all connectors for that disjunct. [10]

Dependency

Connectors may also include head-dependent indicators h and d. In this case, a connector containing a head indicator is only allowed to connect to a connector containing the dependent indicator (or to a connector without any h-d indicators on it). When these indicators are used, the link is decorated with arrows to indicate the link direction. [9]

A recent extension simplifies the specification of connectors for languages that have little or no restrictions on word-order, such as Lithuanian. There are also extensions to make it easier to support languages with concatenative morphologies.

Planarity

The parsing algorithm also requires that the final graph is a planar graph, i.e. that no links cross. [9] This constraint is based on empirical psycho-linguistic evidence that, indeed, for most languages, in nearly all situations, dependency links really do not cross. [11] [12] There are rare exceptions, e.g. in Finnish, and even in English; they can be parsed by link-grammar only by introducing more complex and selective connector types to capture these situations.

Costs and selection

Connectors can have an optional floating-point cost markup, so that some are "cheaper" to use than others, thus giving preference to certain parses over others. [9] That is, the total cost of parse is the sum of the individual costs of the connectors that were used; the cheapest parse indicates the most likely parse. This is used for parse-ranking multiple ambiguous parses. The fact that the costs are local to the connectors, and are not a global property of the algorithm makes them essentially Markovian in nature. [13] [14] [15] [16] [17] [18]

The assignment of a log-likelihood to linkages allows link grammar to implement the semantic selection of predicate-argument relationships. That is, certain constructions, although syntactically valid, are extremely unlikely. In this way, link grammar embodies some of the ideas present in operator grammar.

Because the costs are additive, they behave like the logarithm of the probability (since log-likelihoods are additive), or equivalently, somewhat like the entropy (since entropies are additive). This makes link grammar compatible with machine learning techniques such as hidden Markov models and the Viterbi algorithm, because the link costs correspond to the link weights in Markov networks or Bayesian networks.

Type theory

The link grammar link types can be understood to be the types in the sense of type theory. [9] [19] In effect, the link grammar can be used to model the internal language of certain (non-symmetric) compact closed categories, such as pregroup grammars. In this sense, link grammar appears to be isomorphic or homomorphic to some categorial grammars. Thus, for example, in a categorial grammar the noun phrase "the bad boy" may be written as

whereas the corresponding disjuncts in link grammar would be

the: D+; bad: A+; boy: D- & A-;

The contraction rules (inference rules) of the Lambek calculus can be mapped to the connecting of connectors in link grammar. The + and - directional indicators correspond the forward and backward-slashes of the categorical grammar. Finally, the single-letter names A and D can be understood as labels or "easy-to-read" mnemonic names for the rather more verbose types NP/N, etc.

The primary distinction here is then that the categorical grammars have two type constructors, the forward and backward slashes, that can be used to create new types (such as NP/N) from base types (such as NP and N). Link-grammar omits the use of type constructors, opting instead to define a much larger set of base types having compact, easy-to-remember mnemonics.

Examples

Example 1

A basic rule file for an SVO language might look like:

<determiner>     D+; <noun-subject>  {D−} &  S+; <noun-object>   {D−} &  O−; <verb>           S−  & {O+};

Thus the English sentence, "The boy painted a picture" would appear as:

           +-----O-----+  +-D-+--S--+     +--D--+  |   |     |     |     | The boy painted  a  picture

Similar parses apply for Chinese. [20]

Example 2

Conversely, a rule file for a null subject SOV language might consist of the following links:

<noun-subject>   S+; <noun-object>    O+; <verb>          {O−} & {S−};

And a simple Persian sentence, man nAn xordam (من نان خوردم) 'I ate bread' would look like: [21] [22] [23]

 +-----S-----+  |     +--O--+  |     |     | man   nAn xordam

VSO order can be likewise accommodated, such as for Arabic. [24]

Example 3 (morphology)

In many languages with a concatenative morphology, the stem plays no grammatical role; the grammar is determined by the suffixes. Thus, in Russian, the sentence 'вверху плыли редкие облачка' might have the parse: [25] [26]

    +------------Wd-----------+---------------SIp---------------+     |         +-------EI------+              +--------Api-------+     |         |      +--LLCZD-+       +-LLAQZ+         +--LLCAO-+     |         |      |        |       |      |         |        | LEFT-WALL вверху.e плы.= =ли.vnndpp ре.= =дкие.api облачк.= =а.ndnpi

The subscripts, such as '.vnndpp', are used to indicate the grammatical category. The primary links: Wd, EI, SIp and Api connect together the suffixes, as, in principle, other stems could appear here, without altering the structure of the sentence. The Api link indicates the adjective; SIp denotes subject-verb inversion; EI is a modifier. The Wd link is used to indicate the head noun; the head verb is not indicated in this sentence. The LLXXX links serve only to attach stems to suffixes.

Example 4 (phonology)

The link-grammar can also indicate phonological agreement between neighboring words. For example:

                     +---------Ost--------+     +------>WV------>+   +------Ds**x-----+     +----Wd---+-Ss*b-+   +--PHv-+----A----+     |         |      |   |      |         | LEFT-WALL that.j-p is.v an abstract.a concept.n

Here, the connector 'PH' is used to constrain the determiners that can appear before the word 'abstract'. It effectively blocks (makes it costly) to use the determiner 'a' in this sentence, while the link to 'an' becomes cheap. The other links are roughly as in previous examples: S denoting subject, O denoting object, D denoting determiner. The 'WV' link indicates the head verb, and the 'W' link indicates the head noun. The lower-case letters following the upper-case link types serve to refine the type; so for example, Ds can only connect to a singular noun; Ss only to a singular subject, Os to a singular object. The lower-case v in PHv denotes 'vowel'; the lower-case d in Wd denotes a declarative sentence.

Example 5 (Vietnamese)

The Vietnamese language sentence "Bữa tiệc hôm qua là một thành công lớn" - "The party yesterday was a great success" may be parsed as follows: [27]

Vietnames link grammar example.png

Implementations

Link Grammar parser
Developer(s) OpenCog
Initial releaseOctober 1991;32 years ago (1991-10) [1]
Stable release
5.8.1 / January 8, 2021;3 years ago (2021-01-08) [28]
Repository
Written in C++; originally C
Operating system Cross-platform
Platform GNU
Type NLP
License LGPLv2
Website www.abisource.com/projects/link-grammar/

The link grammar syntax parser is a library for natural language processing written in C. It is available under the LGPL license. The parser [29] is an ongoing project. Recent versions include improved sentence coverage, Russian, Persian and Arabic language support, prototypes for German, Hebrew, Lithuanian, Vietnamese and Turkish, and programming API's for Python, Java, Common LISP, AutoIt and OCaml, with 3rd-party bindings for Perl, [30] Ruby [31] and JavaScript node.js. [32]

A current major undertaking is a project to learn the grammar and morphology of new languages, using unsupervised learning algorithms. [33] [34]

The link-parser program along with rules and word lists for English may be found in standard Linux distributions, e.g., as a Debian package, although many of these are years out of date. [35]

Applications

AbiWord checks grammar using link grammar Abiword grammar.jpg
AbiWord checks grammar using link grammar

AbiWord, [29] a free word processor, uses link grammar for on-the-fly grammar checking. Words that cannot be linked anywhere are underlined in green.

The semantic relationship extractor RelEx, [36] layered on top of the link grammar library, generates a dependency grammar output by making explicit the semantic relationships between words in a sentence. Its output can be classified as being at a level between that of SSyntR and DSyntR of Meaning-Text Theory. It also provides framing/grounding, anaphora resolution, head-word identification, lexical chunking, part-of-speech identification, and tagging, including entity, date, money, gender, etc. tagging. It includes a compatibility mode to generate dependency output compatible with the Stanford parser, [37] and Penn Treebank [38] -compatible POS tagging.

Link grammar has also been employed for information extraction of biomedical texts [39] [40] and events described in news articles, [41] as well as experimental machine translation systems from English to German, Turkish, Indonesian. [42] and Persian. [43] [44]

The link grammar link dictionary is used to generate and verify the syntactic correctness of three different natural language generation systems: NLGen, [45] NLGen2 [46] and microplanner/surreal. [47] It is also used as a part of the NLP pipeline in the OpenCog AI project.

Notes

  1. 1 2 Daniel Sleator (September 8, 2004). "Link Grammar Bibliography". cmu.edu. Retrieved 2023-08-28.
  2. Anssi Yli-Jyrä & Matti Nykänen (2004). "A Hierarchy of Mildly Context-Sensitive Dependency Grammars" (PDF). In G. P. Gerhard Jäger, Paola Monachesi and S. Wintner (ed.). Proceedings of the 9th conference on Formal Grammar 2004 "FGNancy". Pre-Proceedings. pp. 151–165.
  3. Özlem İstek (2006). A Link Grammar for Turkish (PDF) (Master's thesis). Ankara, Turkey: Bilkent University. Retrieved 2023-08-23.
  4. WV Link type
  5. W link type
  6. S link type
  7. P link type
  8. Daniel D. K. Sleator; Davy Temperley (1991). "Parsing English with a Link Grammar". arXiv: cmp-lg/9508004 .
  9. 1 2 3 4 5 An Introduction to the Link Grammar Parser
  10. Dennis Grinberg; John Lafferty; Daniel Sleator (1995). A Robust Parsing Algorithm for Link Grammar (PDF). Proceedings of the Fourth International Workshop on Parsing Technologies, Prague. Retrieved 2023-08-28.
  11. J. Havelka (2007). Beyond projectivity: multilingual evaluation of constraints and measures on non-projective structures. Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics. Prague, Czech Republic: Association for Computational Linguistics. pp. 608–615.
  12. R. Ferrer i Cancho (2006). "Why do syntactic links not cross?". EPL. 76 (6): 1228–1234.
  13. John Lafferty; Daniel Sleator; Davey Temperley (1992). Grammatical Trigrams: a Probabilistic Model of Link Grammar (PDF). Proceedings of the AAAI Conference on Probabilistic Approaches to Natural Language.
  14. _Ramon Ferrer-i-Cancho (2013). "Hubiness, length, crossings and their relationships in dependency trees". arXiv: 1304.4086 .
  15. D. Temperley (2008). "Dependency length minimization in natural and artificial languages". Journal of Quantitative Linguistics. 15 (3): 256–282.
  16. E. Gibson (2000). "The dependency locality theory: A distance-based theory of linguistic complexity". In Marantz, A.; Miyashita, Y.; O'Neil, W. (eds.). Image, Language, Brain: Papers from the first Mind Articulation Project Symposium. Cambridge, MA: MIT Press.
  17. Haitao Liu (2008). "Dependency distance as a metric of language comprehension difficulty" (PDF). Journal of Cognitive Science. 9 (2): 159–191.
  18. Richard Futrell; Kyle Mahowald; Edward Gibson (2015). "Large-scale evidence of dependency length minimization in 37 languages". PNAS. 112 (33): 10336–10341. doi: 10.1073/pnas.1502134112 . PMC   4547262 .
  19. Daniel Sleator; Davey Temperley (1993). Parsing English with a Link Grammar (PDF). Third International Workshop on Parsing Technologies. (See section 6 on categorial grammar).
  20. Carol Liu (2001). "Towards A Link Grammar for Chinese". Computer Processing of Chinese and Oriental Languages. Chinese Language Computer Society.
  21. John Dehdari; Deryle Lonsdale (2005). "A Link Grammar for Persian" (PDF). Ohio-state.edu. Archived from the original (PDF) on 2008-12-03.
  22. Armin Sajadi; A. Abdollahzadeh (2006). "Farsi Syntactic Analysis using Link Grammar" (PDF). Letter of Research Center of Intelligent Signal Processing (in Persian). 1 (9): 25–37. Archived from the original (PDF) on 2014-04-01.
  23. A. Sajadi; M. Homayounpour (2006). "Representation of Farsi Morphological Knowledge using Link Grammar". Letter of Research Center of Intelligent Signal Processing (in Persian). 1 (9): 41–55.
  24. Warren Casbeer; Jon Dehdari; Deryle Lonsdale (March 2006). A Link Grammar parser for Arabic (PDF). Perspectives on Arabic Linguistics: Papers from the annual symposium on Arabic linguistics. Vol. XX. Kalamazoo, Michigan. Archived from the original (PDF) on 2014-05-12.
  25. Документация по связям и по классам слов доступна.
  26. Грамматика связей (Link Grammar)
  27. Nguyễn Thị Thu Hương, Nguyễn Thúc Hải, Nguyễn Thanh Thủy "Parsing complex - compound sentences with an extension of Vietnamese link parser combined with discourse segmenter" Journal of Computer Science and Cybernetics, Vol 28, No 4 (2012)
  28. www.abisource.com/downloads/link-grammar/
  29. Lingua-LinkParser (Perl interfaces)
  30. "Ruby Link Parser interfaces". Archived from the original on 2016-03-04. Retrieved 2019-02-01.
  31. javaScript node.js library
  32. OpenCog Language Learning
  33. Learning Language from a Large (Unannotated) Corpus
  34. Debian - Package Search Results - link-grammar
  35. "RelEx Dependency Relationship Extractor". Archived from the original on 2009-07-28. Retrieved 2013-11-21.
  36. The Stanford Parser: A statistical parser
  37. The Penn Treebank Project Archived 2013-11-09 at the Wayback Machine
  38. Jing Ding; Daniel Berleant; Jun Xu; Andy W. Fulmer (November 2003). "Extracting biochemical interactions from MEDLINE using a link grammar parser" (PDF). Proceedings of the Fifteenth IEEE Conference on Tools with Artificial Intelligence (ICTAI), 2003. pp. 467–471. ISBN   0-7695-2038-3. Archived from the original (PDF) on 2011-03-31. Retrieved 2023-08-27.
  39. Sampo Pyysalo, Tapio Salakoski, Sophie Aubin and Adeline Nazarenko, "Lexical Adaptation of Link Grammar to the Biomedical Sublanguage: a Comparative Evaluation of Three Approaches", BMC Bioinformatics7(Suppl 3):S2 (2006).
  40. Harsha V. Madhyastha; N. Balakrishnan; K. R. Ramakrishnan (2003). "Event Information Extraction Using Link Grammar". 13th International WorkShop on Research Issues in Data Engineering: Multi-lingual Information Management (RIDE'03). p. 16. doi:10.1109/RIDE.2003.1249841.
  41. Teguh Bharata Adji; Baharum Baharudin; Norshuhani Zamin (2008). "Applying Link Grammar Formalism in the Development of English-Indonesian Machine Translation System". Intelligent Computer Mathematics, 9th International Conference, AISC 2008, 15th Symposium, Calculemus 2008, 7th International Conference, Birmingham, UK, Proceedings. pp. 17–23. doi:10.1007/978-3-540-85110-3_3.
  42. A.Sajadi and M.R Borujerdi, "Machine Translation Using Link Grammar", Submitted to the Journal of Computational Linguistics, MIT Press (Feb 2009)
  43. Sajadi, A., Borujerdi, M. "Machine Translation Based on Unification Link Grammar" Journal of Artificial Intelligence Review. DOI=10.1007/s10462-011-9261-7, Pages 109-132, 2013.
  44. Ruiting Lian, et al, "Sentence generation for artificial brains: a glocal similarity matching approach", Neurocomputing (Elsevier) (2009, submitted for publication).
  45. Blake Lemoine, NLGen2: A Linguistically Plausible, General Purpose Natural Language Generation System (2009)
  46. Microplanner and Surface Realization (SuReal)

Language extensions

Related Research Articles

In linguistics, syntax is the study of how words and morphemes combine to form larger units such as phrases and sentences. Central concerns of syntax include word order, grammatical relations, hierarchical sentence structure (constituency), agreement, the nature of crosslinguistic variation, and the relationship between form and meaning (semantics). There are numerous approaches to syntax that differ in their central assumptions and goals.

In grammar, a phrase—called expression in some contexts—is a group of words or singular word acting as a grammatical unit. For instance, the English expression "the very happy squirrel" is a noun phrase which contains the adjective phrase "very happy". Phrases can consist of a single word or a complete sentence. In theoretical linguistics, phrases are often analyzed as units of syntactic structure such as a constituent. There is a difference between the common use of the term phrase and its technical use in linguistics. In common usage, a phrase is usually a group of words with some special idiomatic meaning or other significance, such as "all rights reserved", "economical with the truth", "kick the bucket", and the like. It may be a euphemism, a saying or proverb, a fixed expression, a figure of speech, etc.. In linguistics, these are known as phrasemes.

In grammar, a part of speech or part-of-speech is a category of words that have similar grammatical properties. Words that are assigned to the same part of speech generally display similar syntactic behavior, sometimes similar morphological behavior in that they undergo inflection for similar properties and even similar semantic behavior. Commonly listed English parts of speech are noun, verb, adjective, adverb, pronoun, preposition, conjunction, interjection, numeral, article, and determiner.

English grammar is the set of structural rules of the English language. This includes the structure of words, phrases, clauses, sentences, and whole texts.

A noun phrase – or NP or nominal (phrase) – is a phrase that usually has a noun or pronoun as its head, and has the same grammatical functions as a noun. Noun phrases are very common cross-linguistically, and they may be the most frequently occurring phrase type.

In language, a clause is a constituent that comprises a semantic predicand and a semantic predicate. A typical clause consists of a subject and a syntactic predicate, the latter typically a verb phrase composed of a verb with any objects and other modifiers. However, the subject is sometimes unvoiced if it is retrievable from context, especially in null-subject language but also in other languages, including English instances of the imperative mood.

<span class="mw-page-title-main">Parse tree</span> Tree in formal language theory

A parse tree or parsing tree or derivation tree or concrete syntax tree is an ordered, rooted tree that represents the syntactic structure of a string according to some context-free grammar. The term parse tree itself is used primarily in computational linguistics; in theoretical syntax, the term syntax tree is more common.

In linguistics, an object is any of several types of arguments. In subject-prominent, nominative-accusative languages such as English, a transitive verb typically distinguishes between its subject and any of its objects, which can include but are not limited to direct objects, indirect objects, and arguments of adpositions ; the latter are more accurately termed oblique arguments, thus including other arguments not covered by core grammatical roles, such as those governed by case morphology or relational nouns . In ergative-absolutive languages, for example most Australian Aboriginal languages, the term "subject" is ambiguous, and thus the term "agent" is often used instead to contrast with "object", such that basic word order is often spoken of in terms such as Agent-Object-Verb (AOV) instead of Subject-Object-Verb (SOV). Topic-prominent languages, such as Mandarin, focus their grammars less on the subject-object or agent-object dichotomies but rather on the pragmatic dichotomy of topic and comment.

A language is head-marking if the grammatical marks showing agreement between different words of a phrase tend to be placed on the heads of phrases, rather than on the modifiers or dependents. Many languages employ both head-marking and dependent-marking, and some languages double up and are thus double-marking. The concept of head/dependent-marking was proposed by Johanna Nichols in 1986 and has come to be widely used as a basic category in linguistic typology.

In linguistics, branching refers to the shape of the parse trees that represent the structure of sentences. Assuming that the language is being written or transcribed from left to right, parse trees that grow down and to the right are right-branching, and parse trees that grow down and to the left are left-branching. The direction of branching reflects the position of heads in phrases, and in this regard, right-branching structures are head-initial, whereas left-branching structures are head-final. English has both right-branching (head-initial) and left-branching (head-final) structures, although it is more right-branching than left-branching. Some languages such as Japanese and Turkish are almost fully left-branching (head-final). Some languages are mostly right-branching (head-initial).

Dependency grammar (DG) is a class of modern grammatical theories that are all based on the dependency relation and that can be traced back primarily to the work of Lucien Tesnière. Dependency is the notion that linguistic units, e.g. words, are connected to each other by directed links. The (finite) verb is taken to be the structural center of clause structure. All other syntactic units (words) are either directly or indirectly connected to the verb in terms of the directed links, which are called dependencies. Dependency grammar differs from phrase structure grammar in that while it can identify phrases it tends to overlook phrasal nodes. A dependency structure is determined by the relation between a word and its dependents. Dependency structures are flatter than phrase structures in part because they lack a finite verb phrase constituent, and they are thus well suited for the analysis of languages with free word order, such as Czech or Warlpiri.

A sentence diagram is a pictorial representation of the grammatical structure of a sentence. The term "sentence diagram" is used more when teaching written language, where sentences are diagrammed. The model shows the relations between words and the nature of sentence structure and can be used as a tool to help recognize which potential sentences are actual sentences.

Traditional grammar is a framework for the description of the structure of a language. The roots of traditional grammar are in the work of classical Greek and Latin philologists. The formal study of grammar based on these models became popular during the Renaissance.

<span class="mw-page-title-main">Grammatical relation</span>

In linguistics, grammatical relations are functional relationships between constituents in a clause. The standard examples of grammatical functions from traditional grammar are subject, direct object, and indirect object. In recent times, the syntactic functions, typified by the traditional categories of subject and object, have assumed an important role in linguistic theorizing, within a variety of approaches ranging from generative grammar to functional and cognitive theories. Many modern theories of grammar are likely to acknowledge numerous further types of grammatical relations. The role of grammatical relations in theories of grammar is greatest in dependency grammars, which tend to posit dozens of distinct grammatical relations. Every head-dependent dependency bears a grammatical function.

In linguistics, head directionality is a proposed parameter that classifies languages according to whether they are head-initial or head-final. The head is the element that determines the category of a phrase: for example, in a verb phrase, the head is a verb. Therefore, head initial would be "VO" languages and head final would be "OV" languages.

A definite clause grammar (DCG) is a way of expressing grammar, either for natural or formal languages, in a logic programming language such as Prolog. It is closely related to the concept of attribute grammars / affix grammars. DCGs are usually associated with Prolog, but similar languages such as Mercury also include DCGs. They are called definite clause grammars because they represent a grammar as a set of definite clauses in first-order logic.

Statistical parsing is a group of parsing methods within natural language processing. The methods have in common that they associate grammar rules with a probability. Grammar rules are traditionally viewed in computational linguistics as defining the valid sentences in a language. Within this mindset, the idea of associating each rule with a probability then provides the relative frequency of any given grammar rule and, by deduction, the probability of a complete parse for a sentence. Using this concept, statistical parsers make use of a procedure to search over a space of all candidate parses, and the computation of each candidate's probability, to derive the most probable parse of a sentence. The Viterbi algorithm is one popular method of searching for the most probable parse.

This article describes the grammar of the Scottish Gaelic language.

The grammar of the Klingon language was created by Marc Okrand for the Star Trek franchise. He first described it in his book The Klingon Dictionary. It is a nominative–accusative, primarily suffixing agglutinative language, and has an object–verb–subject word order. The Klingon language has a number of unusual grammatical features, as it was designed to sound and seem alien, but it has an extremely regular morphology.

Universal Dependencies, frequently abbreviated as UD, is an international cooperative project to create treebanks of the world's languages. These treebanks are openly accessible and available. Core applications are automated text processing in the field of natural language processing (NLP) and research into natural language syntax and grammar, especially within linguistic typology. The project's primary aim is to achieve cross-linguistic consistency of annotation, while still permitting language-specific extensions when necessary. The annotation scheme has it roots in three related projects: Stanford Dependencies, Google universal part-of-speech tags, and the Interset interlingua for morphosyntactic tagsets. The UD annotation scheme uses a representation in the form of dependency trees as opposed to a phrase structure trees. At the present time, there are just over 200 treebanks of more than 100 languages available in the UD inventory.