Functional presence engine

Last updated

A Functional Presence Engine, or FPE, is a probabilistic parsing mechanism that uses at least four components to respond to input patterns. It comprises a lexing system, a probabilistic fitness function, a knowledge base, and a library of functions that the knowledge base can trigger.

The lexing system accepts and parses inputs and or query patterns. The probabilistic fitness mechanism determines close approximations and viable responses to the input patterns from a given knowledge base and then selects one or more functions that produce appropriate responses. A Functional Presence Engines is, subsequently, a stimulus-response mechanism that allows for a higher variability of inputs to elicit response patterns with a high likelihood of correctness, even from incomplete training. The system predates SIRI by six years. [1]

Such systems allow conversational AI and virtual assistant platforms [2] to respond correctly to new inputs outside their training sets – The US Army's Sgt Star [3] being a prime example. FPEs are widely used for intelligent customer service systems and for digital assistants. FPEs have also been deployed as black-box solutions and embedded in security appliances. [4]

History

The first Functional Presence Engine was deployed in 2001 by Spectre AI Incorporated. The technology and a number of embodiments were subsequently patented by Spectre AI's cofounder Robert Hust, [5] the FPE's original inventor, and Mark Zartler [6] who was Spectre AI's lead developer.

The development of the FPE also resulted in an obscure markup language that the company referred to as FPML (Functional Presence Markup Language), which was based largely on AIML (Artificial Intelligence Markup Language).

The original FPE and FPML are now proprietary technologies owned by Verint Systems. [7]

Related Research Articles

In computer science, the Earley parser is an algorithm for parsing strings that belong to a given context-free language, though it may suffer problems with certain nullable grammars. The algorithm, named after its inventor, Jay Earley, is a chart parser that uses dynamic programming; it is mainly used for parsing in computational linguistics. It was first introduced in his dissertation in 1968.

Lexical tokenization is conversion of a text into meaningful lexical tokens belonging to categories defined by a "lexer" program. In case of a natural language, those categories include nouns, verbs, adjectives, punctuations etc. In case of a programming language, the categories include identifiers, operators, grouping symbols and data types. Lexical tokenization is not the same process as the probabilistic tokenization, used for large language model's data preprocessing, that encode text into numerical tokens, using byte pair encoding.

<span class="mw-page-title-main">Pattern recognition</span> Automated recognition of patterns and regularities in data

Pattern recognition is the automated recognition of patterns and regularities in data. While similar, pattern recognition (PR) is not to be confused with pattern machines (PM) which may possess (PR) capabilities but their primary function is to distinguish and create emergent pattern. PR has applications in statistical data analysis, signal processing, image analysis, information retrieval, bioinformatics, data compression, computer graphics and machine learning. Pattern recognition has its origins in statistics and engineering; some modern approaches to pattern recognition include the use of machine learning, due to the increased availability of big data and a new abundance of processing power.

<span class="mw-page-title-main">Perceptron</span> Algorithm for supervised learning of binary classifiers

In machine learning, the perceptron is an algorithm for supervised learning of binary classifiers. A binary classifier is a function which can decide whether or not an input, represented by a vector of numbers, belongs to some specific class. It is a type of linear classifier, i.e. a classification algorithm that makes its predictions based on a linear predictor function combining a set of weights with the feature vector.

<span class="mw-page-title-main">Machine learning</span> Study of algorithms that improve automatically through experience

Machine learning (ML) is an umbrella term for solving problems for which development of algorithms by human programmers would be cost-prohibitive, and instead the problems are solved by helping machines "discover" their "own" algorithms, without needing to be explicitly told what to do by any human-developed algorithms. Recently, generative artificial neural networks have been able to surpass results of many previous approaches. Machine-learning approaches have been applied to large language models, computer vision, speech recognition, email filtering, agriculture and medicine, where it is too costly to develop algorithms to perform the needed tasks.

Parsing, syntax analysis, or syntactic analysis is the process of analyzing a string of symbols, either in natural language, computer languages or data structures, conforming to the rules of a formal grammar. The term parsing comes from Latin pars (orationis), meaning part.

In computing, memoization or memoisation is an optimization technique used primarily to speed up computer programs by storing the results of expensive function calls to pure functions and returning the cached result when the same inputs occur again. Memoization has also been used in other contexts, such as in simple mutually recursive descent parsing. It is a type of caching, distinct from other forms of caching such as buffering and page replacement. In the context of some logic programming languages, memoization is also known as tabling.

Natural-language user interface is a type of computer human interface where linguistic phenomena such as verbs, phrases and clauses act as UI controls for creating, selecting and modifying data in software applications.

In science, computing, and engineering, a black box is a system which can be viewed in terms of its inputs and outputs, without any knowledge of its internal workings. Its implementation is "opaque" (black). The term can be used to refer to many inner workings, such as those of a transistor, an engine, an algorithm, the human brain, or an institution or government.

ZK is an open-source Ajax Web application framework, written in Java, that enables creation of graphical user interfaces for Web applications with little required programming knowledge.

<span class="mw-page-title-main">OpenCog</span> Project for an open source artificial intelligence framework

OpenCog is a project that aims to build an open source artificial intelligence framework. OpenCog Prime is an architecture for robot and virtual embodied cognition that defines a set of interacting components designed to give rise to human-equivalent artificial general intelligence (AGI) as an emergent phenomenon of the whole system. OpenCog Prime's design is primarily the work of Ben Goertzel while the OpenCog framework is intended as a generic framework for broad-based AGI research. Research utilizing OpenCog has been published in journals and presented at conferences and workshops including the annual Conference on Artificial General Intelligence. OpenCog is released under the terms of the GNU Affero General Public License.

Statistical language acquisition, a branch of developmental psycholinguistics, studies the process by which humans develop the ability to perceive, produce, comprehend, and communicate with natural language in all of its aspects through the use of general learning mechanisms operating on statistical patterns in the linguistic input. Statistical learning acquisition claims that infants' language-learning is based on pattern perception rather than an innate biological grammar. Several statistical elements such as frequency of words, frequent frames, phonotactic patterns and other regularities provide information on language structure and meaning for facilitation of language acquisition.

The following outline is provided as an overview of and topical guide to natural-language processing:

<span class="mw-page-title-main">Wolfram Language</span> Programming language and environment developed by Wolfram Research

The Wolfram Language is a proprietary, general high-level multi-paradigm programming language developed by Wolfram Research. It emphasizes symbolic computation, functional programming, and rule-based programming and can employ arbitrary structures and data. It is the programming language of the mathematical symbolic computation program Mathematica.

Inductive programming (IP) is a special area of automatic programming, covering research from artificial intelligence and programming, which addresses learning of typically declarative and often recursive programs from incomplete specifications, such as input/output examples or constraints.

Mycroft is a free and open-source software virtual assistant that uses a natural language user interface. Its code was formerly copyleft, but is now under a permissive license. It is named after a fictional computer from 1966 science fiction novel The Moon Is a Harsh Mistress.

<span class="mw-page-title-main">Glossary of artificial intelligence</span> List of definitions of terms and concepts commonly used in the study of artificial intelligence

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

Spectre AI Incorporated was a private software company that served various government agencies and defense contractors in the early 2000s. The company is notable for having developed and deployed the first functional presence engine in 2001. Spectre AI's initial contract with Raytheon was the direct result of a request on the floor of the United States Congress on September 13, 2001 by Senators Judd Gregg and Ernest Hollings. This was a surprising amount of access given the company was started one year earlier in Spokane, Washington.

References

  1. "The History of Apple's Siri". SRI International. Retrieved 2021-03-03.
  2. "Next IT Debuts Alme Virtual Assistant Healthcare Platform". Speech Technology Magazine. 2013-10-18. Retrieved 2021-03-03.
  3. EFF, David Maass- (18 April 2014). "Everything We Know About the Army's Uncanny Chatbots". Gizmodo. Retrieved 2021-03-03.
  4. "System and method for providing network support services and premises gateway support infrastructure". Google Patents.
  5. "Robert Hust, Patents". patents.justia.com.
  6. "Mark Zartler, Patents". patents.justia.com.
  7. "M&A Watch: Verint Acquires Next IT, Accelerating Acceptance of Omnichannel Intelligent Assistance |". 2017-12-21. Retrieved 2021-03-03.