Semantic decomposition (natural language processing)

Last updated

A semantic decomposition is an algorithm that breaks down the meanings of phrases or concepts into less complex concepts. [1] The result of a semantic decomposition is a representation of meaning. [2] This representation can be used for tasks, such as those related to artificial intelligence or machine learning. Semantic decomposition is common in natural language processing applications.

Contents

The basic idea of a semantic decomposition is taken from the learning skills of adult humans, where words are explained using other words. It is based on Meaning-text theory. Meaning-text theory is used as a theoretical linguistic framework to describe the meaning of concepts with other concepts.

Background

Given that an AI does not inherently have language, it is unable to think about the meanings behind the words of a language. An artificial notion of meaning needs to be created for a strong AI to emerge. [3]

Creating an artificial representation of meaning requires the analysis of what meaning is. Many terms are associated with meaning, including semantics, pragmatics, knowledge and understanding or word sense. [4] Each term describes a particular aspect of meaning, and contributes to a multitude of theories explaining what meaning is. These theories need to be analyzed further to develop an artificial notion of meaning best fit for our current state of knowledge.

Graph representations

Abstract approach on how knowledge representation and reasoning allow a problem specific solution (answer) to a given problem (questions) Knowledge Reasoning.pdf
Abstract approach on how knowledge representation and reasoning allow a problem specific solution (answer) to a given problem (questions)

Representing meaning as a graph is one of the two ways that both an AI cognition and a linguistic researcher think about meaning (connectionist view). Logicians utilize a formal representation of meaning to build upon the idea of symbolic representation, whereas description logics describe languages and the meaning of symbols. This contention between 'neat' and 'scruffy' techniques has been discussed since the 1970s. [5]

Research has so far identified semantic measures and with that word-sense disambiguation (WSD) - the differentiation of meaning of words - as the main problem of language understanding. [6] As an AI-complete environment, WSD is a core problem of natural language understanding. [7] [8] AI approaches that use knowledge-given reasoning creates a notion of meaning combining the state of the art knowledge of natural meaning with the symbolic and connectionist formalization of meaning for AI. The abstract approach is shown in Figure. First, a connectionist knowledge representation is created as a semantic network consisting of concepts and their relations to serve as the basis for the representation of meaning. [9] [10] [11] [12]

This graph is built out of different knowledge sources like WordNet, Wiktionary, and BabelNET. The graph is created by lexical decomposition that recursively breaks each concept semantically down into a set of semantic primes. [1] The primes are taken from the theory of Natural Semantic Metalanguage, [13] which has been analyzed for usefulness in formal languages. [14] Upon this graph marker passing [15] [16] [17] is used to create the dynamic part of meaning representing thoughts. [18] The marker passing algorithm, where symbolic information is passed along relations form one concept to another, uses node and edge interpretation to guide its markers. The node and edge interpretation model is the symbolic influence of certain concepts.

Future work uses the created representation of meaning to build heuristics and evaluate them through capability matching and agent planning, chatbots or other applications of natural language understanding.

See also

Related Research Articles

Knowledge representation and reasoning is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can use to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language. Knowledge representation incorporates findings from psychology about how humans solve problems and represent knowledge, in order to design formalisms that will make complex systems easier to design and build. Knowledge representation and reasoning also incorporates findings from logic to automate various kinds of reasoning.

Natural language processing (NLP) is an interdisciplinary subfield of computer science and artificial intelligence. It is primarily concerned with providing computers the ability to process data encoded in natural language and is thus closely related to information retrieval, knowledge representation and computational linguistics, a subfield of linguistics. Typically data is collected in text corpora, using either rule-based, statistical or neural-based approaches of machine learning and deep learning.

<span class="mw-page-title-main">Semantic network</span> Knowledge base that represents semantic relations between concepts in a network

A semantic network, or frame network is a knowledge base that represents semantic relations between concepts in a network. This is often used as a form of knowledge representation. It is a directed or undirected graph consisting of vertices, which represent concepts, and edges, which represent semantic relations between concepts, mapping or connecting semantic fields. A semantic network may be instantiated as, for example, a graph database or a concept map. Typical standardized semantic networks are expressed as semantic triples.

Word-sense disambiguation is the process of identifying which sense of a word is meant in a sentence or other segment of context. In human language processing and cognition, it is usually subconscious.

Natural semantic metalanguage (NSM) is a linguistic theory that reduces lexicons down to a set of semantic primitives. It is based on the conception of Polish professor Andrzej Bogusławski. The theory was formally developed by Anna Wierzbicka at Warsaw University and later at the Australian National University in the early 1970s, and Cliff Goddard at Australia's Griffith University.

<span class="mw-page-title-main">Symbolic artificial intelligence</span> Methods in artificial intelligence research

In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.

The language of thought hypothesis (LOTH), sometimes known as thought ordered mental expression (TOME), is a view in linguistics, philosophy of mind and cognitive science, forwarded by American philosopher Jerry Fodor. It describes the nature of thought as possessing "language-like" or compositional structure. On this view, simple concepts combine in systematic ways to build thoughts. In its most basic form, the theory states that thought, like language, has syntax.

Conceptual semantics is a framework for semantic analysis developed mainly by Ray Jackendoff in 1976. Its aim is to provide a characterization of the conceptual elements by which a person understands words and sentences, and thus to provide an explanatory semantic representation. Explanatory in this sense refers to the ability of a given linguistic theory to describe how a component of language is acquired by a child.

Semantic similarity is a metric defined over a set of documents or terms, where the idea of distance between items is based on the likeness of their meaning or semantic content as opposed to lexicographical similarity. These are mathematical tools used to estimate the strength of the semantic relationship between units of language, concepts or instances, through a numerical description obtained according to the comparison of information supporting their meaning or describing their nature. The term semantic similarity is often confused with semantic relatedness. Semantic relatedness includes any relation between two terms, while semantic similarity only includes "is a" relations. For example, "car" is similar to "bus", but is also related to "road" and "driving".

Cognitive semantics is part of the cognitive linguistics movement. Semantics is the study of linguistic meaning. Cognitive semantics holds that language is part of a more general human cognitive ability, and can therefore only describe the world as people conceive of it. It is implicit that different linguistic communities conceive of simple things and processes in the world differently, not necessarily some difference between a person's conceptual world and the real world.

Frame semantics is a theory of linguistic meaning developed by Charles J. Fillmore that extends his earlier case grammar. It relates linguistic semantics to encyclopedic knowledge. The basic idea is that one cannot understand the meaning of a single word without access to all the essential knowledge that relates to that word. For example, one would not be able to understand the word "sell" without knowing anything about the situation of commercial transfer, which also involves, among other things, a seller, a buyer, goods, money, the relation between the money and the goods, the relations between the seller and the goods and the money, the relation between the buyer and the goods and the money and so on. Thus, a word activates, or evokes, a frame of semantic knowledge relating to the specific concept to which it refers.

<span class="mw-page-title-main">Distributional semantics</span> Field of linguistics

Distributional semantics is a research area that develops and studies theories and methods for quantifying and categorizing semantic similarities between linguistic items based on their distributional properties in large samples of language data. The basic idea of distributional semantics can be summed up in the so-called distributional hypothesis: linguistic items with similar distributions have similar meanings.

Spreading activation is a method for searching associative networks, biological and artificial neural networks, or semantic networks. The search process is initiated by labeling a set of source nodes with weights or "activation" and then iteratively propagating or "spreading" that activation out to other nodes linked to the source nodes. Most often these "weights" are real values that decay as activation propagates through the network. When the weights are discrete this process is often referred to as marker passing. Activation may originate from alternate paths, identified by distinct markers, and terminate when two alternate paths reach the same node. However brain studies show that several different brain areas play an important role in semantic processing.

DOGMA, short for Developing Ontology-Grounded Methods and Applications, is the name of research project in progress at Vrije Universiteit Brussel's STARLab, Semantics Technology and Applications Research Laboratory. It is an internally funded project, concerned with the more general aspects of extracting, storing, representing and browsing information.

The linguistics wars were extended disputes among American theoretical linguists that occurred mostly during the 1960s and 1970s, stemming from a disagreement between Noam Chomsky and several of his associates and students. The debates started in 1967 when linguists Paul Postal, John R. Ross, George Lakoff, and James D. McCawley —self-dubbed the "Four Horsemen of the Apocalypse"—proposed an alternative approach in which the relation between semantics and syntax is viewed differently, which treated deep structures as meaning rather than syntactic objects. While Chomsky and other generative grammarians argued that meaning is driven by an underlying syntax, generative semanticists posited that syntax is shaped by an underlying meaning. This intellectual divergence led to two competing frameworks in generative semantics and interpretive semantics.

Meaning–text theory (MTT) is a theoretical linguistic framework, first put forward in Moscow by Aleksandr Žolkovskij and Igor Mel’čuk, for the construction of models of natural language. The theory provides a large and elaborate basis for linguistic description and, due to its formal character, lends itself particularly well to computer applications, including machine translation, phraseology, and lexicography.

In computational linguistics, word-sense induction (WSI) or discrimination is an open problem of natural language processing, which concerns the automatic identification of the senses of a word. Given that the output of word-sense induction is a set of senses for the target word, this task is strictly related to that of word-sense disambiguation (WSD), which relies on a predefined sense inventory and aims to solve the ambiguity of words in context.

SemEval is an ongoing series of evaluations of computational semantic analysis systems; it evolved from the Senseval word sense evaluation series. The evaluations are intended to explore the nature of meaning in language. While meaning is intuitive to humans, transferring those intuitions to computational analysis has proved elusive.

The following outline is provided as an overview of and topical guide to natural-language processing:

In natural language processing (NLP), a word embedding is a representation of a word. The embedding is used in text analysis. Typically, the representation is a real-valued vector that encodes the meaning of the word in such a way that the words that are closer in the vector space are expected to be similar in meaning. Word embeddings can be obtained using language modeling and feature learning techniques, where words or phrases from the vocabulary are mapped to vectors of real numbers.

References

  1. 1 2 Riemer, Nick (2015-07-30). The Routledge Handbook of Semantics. Routledge. ISBN   9781317412441.
  2. Fähndrich, J. (2018). Semantic decomposition and marker passing in an artificial representation of meaning. Technische Universitaet Berlin (Germany).
  3. Loizos Michael. 2015. Jumping to conclusions. In Proceedings of the 2015 International Conference on Defeasible and Ampliative Reasoning - Volume 1423 (DARe'15). CEUR-WS.org, Aachen, DEU, 43–49.
  4. Löbner, Sebastian (2015-05-19). Semantik: Eine Einführung (in German). Walter de Gruyter GmbH & Co KG. ISBN   9783110350906.
  5. Minsky, Marvin L. (1991-06-15). "Logical Versus Analogical or Symbolic Versus Connectionist or Neat Versus Scruffy". AI Magazine. 12 (2): 34. doi:10.1609/aimag.v12i2.894. ISSN   2371-9621.
  6. Word Sense Disambiguation - Algorithms and Applications | Eneko Agirre | Springer.
  7. Nancy Ide and Jean Veronis. Introduction to the special issue on word sense disambiguation: the state of the art. Computational Linguistics, 24(1):2-40, 1998
  8. Yampolskiy, R. V. (2012, April). AI-complete, AI-hard, or AI-easy–classification of problems in AI. In The 23rd Midwest Artificial Intelligence and Cognitive Science Conference, Cincinnati, OH, USA.
  9. Sycara, Katia; Klusch, Matthias; Widoff, Seth; Lu, Jianguo (1999-03-01). "Dynamic service matchmaking among agents in open information environments". ACM SIGMOD Record. 28 (1): 47–53. CiteSeerX   10.1.1.44.914 . doi:10.1145/309844.309895. ISSN   0163-5808. S2CID   10197051.
  10. Oaks, Phillipa; ter Hofstede, Arthur H. M.; Edmond, David (2003), "Capabilities: Describing What Services Can do", Service-Oriented Computing - ICSOC 2003, Lecture Notes in Computer Science, vol. 2910, Springer Berlin Heidelberg, pp. 1–16, CiteSeerX   10.1.1.473.5321 , doi:10.1007/978-3-540-24593-3_1, ISBN   9783540206811, S2CID   11524526
  11. Johannes Fähndrich est First Search Planning of Service Composition Using Incrementally Redefined Context-Dependent Heuristics. In the German Conference Multiagent System Technologies, pages 404-407, Springer Berlin Heidelberg, 2013
  12. Fähndrich, Johannes; Ahrndt, Sebastian; Albayrak, Sahin (2013), "Towards Self-Explaining Agents", Trends in Practical Applications of Agents and Multiagent Systems, Advances in Intelligent Systems and Computing, vol. 221, Springer International Publishing, pp. 147–154, doi:10.1007/978-3-319-00563-8_18, ISBN   9783319005621
  13. Goddard, Cliff; Wierzbicka, Anna, eds. (1994). Semantic and Lexical Universals: Theory and empirical findings. Amsterdam: Benjamins.
  14. Fähndrich, Johannes; Ahrndt, Sebastian; Albayrak, Sahin (2014-10-15). "Formal Language Decomposition into Semantic Primes". Advances in Distributed Computing and Artificial Intelligence Journal . 3 (1): 56–73. doi: 10.14201/ADCAIJ2014385673 . ISSN   2255-2863.
  15. "integrating Marker Passing and Problem Solving: A Spreading Activation Approach To Improved Choice in Planning". CRC Press. 1987-11-01. Retrieved 2018-11-30.
  16. Hirst, Graeme (1987-01-01). Semantic interpretation and the resolution of ambiguity. Cambridge University Press. ISBN   978-0521322034.
  17. "Self-Explanation through Semantic Annotation: A Survey". ResearchGate. Retrieved 2018-11-30.
  18. Crestani, Fabio (1997). "Application of Spreading Activation Techniques in Information Retrieval". Artificial Intelligence Review. 11 (6): 453–482. doi:10.1023/A:1006569829653. S2CID   14668203.