In artificial intelligence (AI), commonsense reasoning is a human-like ability to make presumptions about the type and essence of ordinary situations humans encounter every day. These assumptions include judgments about the nature of physical objects, taxonomic properties, and peoples' intentions. A device that exhibits commonsense reasoning might be capable of drawing conclusions that are similar to humans' folk psychology (humans' innate ability to reason about people's behavior and intentions) and naive physics (humans' natural understanding of the physical world). [1]
Some definitions and characterizations of common sense from different authors include:
NYU professor Ernest Davis characterizes commonsense knowledge as "what a typical seven year old knows about the world", including physical objects, substances, plants, animals, and human society. It usually excludes book-learning, specialized knowledge, and knowledge of conventions; but it sometimes includes knowledge about those topics. For example, knowing how to play cards is specialized knowledge, not "commonsense knowledge"; but knowing that people play cards for fun does count as "commonsense knowledge". [7]
Compared with humans, existing AI lacks several features of human commonsense reasoning; most notably, humans have powerful mechanisms for reasoning about "naïve physics" such as space, time, and physical interactions. This enables even young children to easily make inferences like "If I roll this pen off a table, it will fall on the floor". Humans also have a powerful mechanism of "folk psychology" that helps them to interpret natural-language sentences such as "The city councilmen refused the demonstrators a permit because they advocated violence". (A generic AI has difficulty discerning whether the ones alleged to be advocating violence are the councilmen or the demonstrators.) [1] [8] [9] This lack of "common knowledge" means that AI often makes different mistakes than humans make, in ways that can seem incomprehensible. For example, existing self-driving cars cannot reason about the location nor the intentions of pedestrians in the exact way that humans do, and instead must use non-human modes of reasoning to avoid accidents. [10] [11] [12]
Overlapping subtopics of commonsense reasoning include quantities and measurements, time and space, physics, minds, society, plans and goals, and actions and change. [13]
The commonsense knowledge problem is a current project in the sphere of artificial intelligence to create a database that contains the general knowledge most individuals are expected to have, represented in an accessible way to artificial intelligence programs [14] that use natural language. Due to the broad scope of the commonsense knowledge, this issue is considered to be among the most difficult problems in AI research. [15] In order for any task to be done as a human mind would manage it, the machine is required to appear as intelligent as a human being. Such tasks include object recognition, machine translation and text mining. To perform them, the machine has to be aware of the same concepts that an individual, who possess commonsense knowledge, recognizes.
In 1961, Bar Hillel first discussed the need and significance of practical knowledge for natural language processing in the context of machine translation. [16] Some ambiguities are resolved by using simple and easy to acquire rules. Others require a broad acknowledgement of the surrounding world, thus they require more commonsense knowledge. For instance, when a machine is used to translate a text, problems of ambiguity arise, which could be easily resolved by attaining a concrete and true understanding of the context. Online translators often resolve ambiguities using analogous or similar words. For example, in translating the sentences "The electrician is working" and "The telephone is working" into German, the machine translates correctly "working" in the means of "laboring" in the first one and as "functioning properly" in the second one. The machine has seen and read in the body of texts that the German words for "laboring" and "electrician" are frequently used in a combination and are found close together. The same applies for "telephone" and "function properly". However, the statistical proxy which works in simple cases often fails in complex ones. Existing computer programs carry out simple language tasks by manipulating short phrases or separate words, but they don't attempt any deeper understanding and focus on short-term results.
Issues of this kind arise in computer vision. [1] [17] For instance when looking at a photograph of a bathroom some items that are small and only partly seen, such as facecloths and bottles, are recognizable due to the surrounding objects (toilet, wash basin, bathtub), which suggest the purpose of the room. In an isolated image they would be difficult to identify. Movies prove to be even more difficult tasks. Some movies contain scenes and moments that cannot be understood by simply matching memorized templates to images. For instance, to understand the context of the movie, the viewer is required to make inferences about characters’ intentions and make presumptions depending on their behavior. In the contemporary state of the art, it is impossible to build and manage a program that will perform such tasks as reasoning, i.e. predicting characters’ actions. The most that can be done is to identify basic actions and track characters.
The need and importance of commonsense reasoning in autonomous robots that work in a real-life uncontrolled environment is evident. For instance, if a robot is programmed to perform the tasks of a waiter at a cocktail party, and it sees that the glass he had picked up is broken, the waiter-robot should not pour the liquid into the glass, but instead pick up another one. Such tasks seem obvious when an individual possesses simple commonsense reasoning, but to ensure that a robot will avoid such mistakes is challenging. [1]
Significant progress in the field of the automated commonsense reasoning is made in the areas of the taxonomic reasoning, actions and change reasoning, reasoning about time. Each of these spheres has a well-acknowledged theory for wide range of commonsense inferences. [18]
Taxonomy is the collection of individuals and categories and their relations. Three basic relations are:
Transitivity is one type of inference in taxonomy. Since Tweety is an instance of robin and robin is a subset of bird, it follows that Tweety is an instance of bird. Inheritance is another type of inference. Since Tweety is an instance of robin, which is a subset of bird and bird is marked with property canfly, it follows that Tweety and robin have property canfly. When an individual taxonomizes more abstract categories, outlining and delimiting specific categories becomes more problematic. Simple taxonomic structures are frequently used in AI programs. For instance, WordNet is a resource including a taxonomy, whose elements are meanings of English words. Web mining systems used to collect commonsense knowledge from Web documents focus on taxonomic relations and specifically in gathering taxonomic relations. [1]
The theory of action, events and change is another range of the commonsense reasoning. [19] There are established reasoning methods for domains that satisfy the constraints listed below:
Temporal reasoning is the ability to make presumptions about humans' knowledge of times, durations and time intervals. For example, if an individual knows that Mozart was born after Haydn and died earlier than him, they can use their temporal reasoning knowledge to deduce that Mozart had died younger than Haydn. The inferences involved reduce themselves to solving systems of linear inequalities. [20] To integrate that kind of reasoning with concrete purposes, such as natural language interpretation, is more challenging, because natural language expressions have context dependent interpretation. [21] Simple tasks such as assigning timestamps to procedures cannot be done with total accuracy.
Qualitative reasoning [22] is the form of commonsense reasoning analyzed with certain success. It is concerned with the direction of change in interrelated quantities. For instance, if the price of a stock goes up, the amount of stocks that are going to be sold will go down. If some ecosystem contains wolves and lambs and the number of wolves decreases, the death rate of the lambs will go down as well. This theory was firstly formulated by Johan de Kleer, who analyzed an object moving on a roller coaster. The theory of qualitative reasoning is applied in many spheres such as physics, biology, engineering, ecology, etc. It serves as the basis for many practical programs, analogical mapping, text understanding.
As of 2014, there are some commercial systems trying to make the use of commonsense reasoning significant. However, they use statistical information as a proxy for commonsense knowledge, where reasoning is absent. Current programs manipulate individual words, but they don't attempt or offer further understanding. According to Ernest Davis and Gary Marcus, five major obstacles interfere with the producing of a satisfactory "commonsense reasoner". [1]
Compared with humans, as of 2018 existing computer programs perform extremely poorly on modern "commonsense reasoning" benchmark tests such as the Winograd Schema Challenge. [23] The problem of attaining human-level competency at "commonsense knowledge" tasks is considered to probably be "AI complete" (that is, solving it would require the ability to synthesize a human-level intelligence). [24] [25] Some researchers believe that supervised learning data is insufficient to produce an artificial general intelligence capable of commonsense reasoning, and have therefore turned to less-supervised learning techniques. [26]
Commonsense's reasoning study is divided into knowledge-based approaches and approaches that are based on machine learning over and using a large data corpora with limited interactions between these two types of approaches [ citation needed ]. There are also crowdsourcing approaches, attempting to construct a knowledge basis by linking the collective knowledge and the input of non-expert people. Knowledge-based approaches can be separated into approaches based on mathematical logic [ citation needed ].
In knowledge-based approaches, the experts are analyzing the characteristics of the inferences that are required to do reasoning in a specific area or for a certain task. The knowledge-based approaches consist of mathematically grounded approaches, informal knowledge-based approaches and large-scale approaches. The mathematically grounded approaches are purely theoretical and the result is a printed paper instead of a program. The work is limited to the range of the domains and the reasoning techniques that are being reflected on. In informal knowledge-based approaches, theories of reasoning are based on anecdotal data and intuition that are results from empirical behavioral psychology. Informal approaches are common in computer programming. Two other popular techniques for extracting commonsense knowledge from Web documents involve Web mining and Crowd sourcing.
COMET (2019), which uses both the OpenAI GPT language model architecture and existing commonsense knowledge bases such as ConceptNet, claims to generate commonsense inferences at a level approaching human benchmarks. Like many other current efforts, COMET over-relies on surface language patterns and is judged to lack deep human-level understanding of many commonsense concepts. Other language-model approaches include training on visual scenes rather than just text, and training on textual descriptions of scenarios involving commonsense physics. [6] [27]
Artificial intelligence (AI), in its broadest sense, is intelligence exhibited by machines, particularly computer systems. It is a field of research in computer science that develops and studies methods and software that enable machines to perceive their environment and use learning and intelligence to take actions that maximize their chances of achieving defined goals. Such machines may be called AIs.
In the field of artificial intelligence (AI), tasks that are hypothesized to require artificial general intelligence to solve are informally known as AI-complete or AI-hard. Calling a problem AI-complete reflects the belief that it cannot be solved by a simple specific algorithm.
Cyc is a long-term artificial intelligence project that aims to assemble a comprehensive ontology and knowledge base that spans the basic concepts and rules about how the world works.
In artificial intelligence (AI), an expert system is a computer system emulating the decision-making ability of a human expert. Expert systems are designed to solve complex problems by reasoning through bodies of knowledge, represented mainly as if–then rules rather than through conventional procedural programming code. Expert systems were first created in the 1970s and then proliferated in the 1980s. Expert systems were among the first truly successful forms of AI software. An expert system is divided into two subsystems: 1) a knowledge base, which represents facts and rules; and 2) an inference engine, which applies the rules to the known facts to deduce new facts, and can include explaining and debugging abilities.
Knowledge representation and reasoning is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can use to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language. Knowledge representation incorporates findings from psychology about how humans solve problems and represent knowledge, in order to design formalisms that will make complex systems easier to design and build. Knowledge representation and reasoning also incorporates findings from logic to automate various kinds of reasoning.
Natural language understanding (NLU) or natural language interpretation (NLI) is a subset of natural language processing in artificial intelligence that deals with machine reading comprehension. NLU has been considered an AI-hard problem.
In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.
In the history of artificial intelligence, neat and scruffy are two contrasting approaches to artificial intelligence (AI) research. The distinction was made in the 1970s and was a subject of discussion until the mid-1980s.
Legal informatics is an area within information science.
In artificial intelligence research, commonsense knowledge consists of facts about the everyday world, such as "Lemons are sour", or "Cows say moo", that all humans are expected to know. It is currently an unsolved problem in Artificial General Intelligence. The first AI program to address common sense knowledge was Advice Taker in 1959 by John McCarthy.
Deborah Louise McGuinness is an American computer scientist and researcher at Rensselaer Polytechnic Institute (RPI). She is a professor of Computer, Cognitive and Web Sciences, Industrial and Systems Engineering, and an endowed chair in the Tetherless World Constellation, a multidisciplinary research institution within RPI that focuses on the study of theories, methods and applications of the World Wide Web. Her fields of expertise include interdisciplinary data integration, artificial intelligence, specifically in knowledge representation and reasoning, description logics, the semantic web, explanation, and trust.
The history of artificial intelligence (AI) began in antiquity, with myths, stories and rumors of artificial beings endowed with intelligence or consciousness by master craftsmen. The seeds of modern AI were planted by philosophers who attempted to describe the process of human thinking as the mechanical manipulation of symbols. This work culminated in the invention of the programmable digital computer in the 1940s, a machine based on the abstract essence of mathematical reasoning. This device and the ideas behind it inspired a handful of scientists to begin seriously discussing the possibility of building an electronic brain.
The philosophy of artificial intelligence is a branch of the philosophy of mind and the philosophy of computer science that explores artificial intelligence and its implications for knowledge and understanding of intelligence, ethics, consciousness, epistemology, and free will. Furthermore, the technology is concerned with the creation of artificial animals or artificial people so the discipline is of considerable interest to philosophers. These factors contributed to the emergence of the philosophy of artificial intelligence.
The following outline is provided as an overview of and topical guide to artificial intelligence:
Frames are an artificial intelligence data structure used to divide knowledge into substructures by representing "stereotyped situations". They were proposed by Marvin Minsky in his 1974 article "A Framework for Representing Knowledge". Frames are the primary data structure used in artificial intelligence frame languages; they are stored as ontologies of sets.
James Frederick Allen is an American computational linguist recognized for his contributions to temporal logic, in particular Allen's interval algebra. He is interested in knowledge representation, commonsense reasoning, and natural language understanding, believing that "deep language understanding can only currently be achieved by significant hand-engineering of semantically-rich formalisms coupled with statistical preferences". He is the John H. Dessaurer Professor of Computer Science at the University of Rochester
Knowledge retrieval seeks to return information in a structured form, consistent with human cognitive processes as opposed to simple lists of data items. It draws on a range of fields including epistemology, cognitive psychology, cognitive neuroscience, logic and inference, machine learning and knowledge discovery, linguistics, and information technology.
A legal expert system is a domain-specific expert system that uses artificial intelligence to emulate the decision-making abilities of a human expert in the field of law. Legal expert systems employ a rule base or knowledge base and an inference engine to accumulate, reference and produce expert knowledge on specific subjects within the legal domain.
The Winograd schema challenge (WSC) is a test of machine intelligence proposed in 2012 by Hector Levesque, a computer scientist at the University of Toronto. Designed to be an improvement on the Turing test, it is a multiple-choice test that employs questions of a very specific structure: they are instances of what are called Winograd schemas, named after Terry Winograd, professor of computer science at Stanford University.
This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.
{{cite journal}}
: Cite journal requires |journal=
(help)