Commonsense reasoning

Last updated

In artificial intelligence (AI), commonsense reasoning is a human-like ability to make presumptions about the type and essence of ordinary situations humans encounter every day. These assumptions include judgments about the nature of physical objects, taxonomic properties, and peoples' intentions. A device that exhibits commonsense reasoning might be capable of drawing conclusions that are similar to humans' folk psychology (humans' innate ability to reason about people's behavior and intentions) and naive physics (humans' natural understanding of the physical world). [1]

Contents

Definitions and characterizations

Some definitions and characterizations of common sense from different authors include:

NYU professor Ernest Davis characterizes commonsense knowledge as "what a typical seven year old knows about the world", including physical objects, substances, plants, animals, and human society. It usually excludes book-learning, specialized knowledge, and knowledge of conventions; but it sometimes includes knowledge about those topics. For example, knowing how to play cards is specialized knowledge, not "commonsense knowledge"; but knowing that people play cards for fun does count as "commonsense knowledge". [7]

Commonsense reasoning problem

A self-driving car system may use a neural network to determine which parts of the picture seem to match previous training images of pedestrians, and then model those areas as slow-moving but somewhat unpredictable rectangular prisms that must be avoided. Detection de personne - exemple 3.jpg
A self-driving car system may use a neural network to determine which parts of the picture seem to match previous training images of pedestrians, and then model those areas as slow-moving but somewhat unpredictable rectangular prisms that must be avoided.

Compared with humans, existing AI lacks several features of human commonsense reasoning; most notably, humans have powerful mechanisms for reasoning about "naïve physics" such as space, time, and physical interactions. This enables even young children to easily make inferences like "If I roll this pen off a table, it will fall on the floor". Humans also have a powerful mechanism of "folk psychology" that helps them to interpret natural-language sentences such as "The city councilmen refused the demonstrators a permit because they advocated violence". (A generic AI has difficulty discerning whether the ones alleged to be advocating violence are the councilmen or the demonstrators.) [1] [8] [9] This lack of "common knowledge" means that AI often makes different mistakes than humans make, in ways that can seem incomprehensible. For example, existing self-driving cars cannot reason about the location nor the intentions of pedestrians in the exact way that humans do, and instead must use non-human modes of reasoning to avoid accidents. [10] [11] [12]

Overlapping subtopics of commonsense reasoning include quantities and measurements, time and space, physics, minds, society, plans and goals, and actions and change. [13]

Commonsense knowledge problem

The commonsense knowledge problem is a current project in the sphere of artificial intelligence to create a database that contains the general knowledge most individuals are expected to have, represented in an accessible way to artificial intelligence programs [14] that use natural language. Due to the broad scope of the commonsense knowledge, this issue is considered to be among the most difficult problems in AI research. [15] In order for any task to be done as a human mind would manage it, the machine is required to appear as intelligent as a human being. Such tasks include object recognition, machine translation and text mining. To perform them, the machine has to be aware of the same concepts that an individual, who possess commonsense knowledge, recognizes.

Commonsense in intelligent tasks

In 1961, Bar Hillel first discussed the need and significance of practical knowledge for natural language processing in the context of machine translation. [16] Some ambiguities are resolved by using simple and easy to acquire rules. Others require a broad acknowledgement of the surrounding world, thus they require more commonsense knowledge. For instance, when a machine is used to translate a text, problems of ambiguity arise, which could be easily resolved by attaining a concrete and true understanding of the context. Online translators often resolve ambiguities using analogous or similar words. For example, in translating the sentences "The electrician is working" and "The telephone is working" into German, the machine translates correctly "working" in the means of "laboring" in the first one and as "functioning properly" in the second one. The machine has seen and read in the body of texts that the German words for "laboring" and "electrician" are frequently used in a combination and are found close together. The same applies for "telephone" and "function properly". However, the statistical proxy which works in simple cases often fails in complex ones. Existing computer programs carry out simple language tasks by manipulating short phrases or separate words, but they don't attempt any deeper understanding and focus on short-term results.

Computer vision

Issues of this kind arise in computer vision. [1] [17] For instance when looking at a photograph of a bathroom some items that are small and only partly seen, such as facecloths and bottles, are recognizable due to the surrounding objects (toilet, wash basin, bathtub), which suggest the purpose of the room. In an isolated image they would be difficult to identify. Movies prove to be even more difficult tasks. Some movies contain scenes and moments that cannot be understood by simply matching memorized templates to images. For instance, to understand the context of the movie, the viewer is required to make inferences about characters’ intentions and make presumptions depending on their behavior. In the contemporary state of the art, it is impossible to build and manage a program that will perform such tasks as reasoning, i.e. predicting characters’ actions. The most that can be done is to identify basic actions and track characters.

Robotic manipulation

The need and importance of commonsense reasoning in autonomous robots that work in a real-life uncontrolled environment is evident. For instance, if a robot is programmed to perform the tasks of a waiter at a cocktail party, and it sees that the glass he had picked up is broken, the waiter-robot should not pour the liquid into the glass, but instead pick up another one. Such tasks seem obvious when an individual possesses simple commonsense reasoning, but to ensure that a robot will avoid such mistakes is challenging. [1]

Successes in automated commonsense reasoning

Significant progress in the field of the automated commonsense reasoning is made in the areas of the taxonomic reasoning, actions and change reasoning, reasoning about time. Each of these spheres has a well-acknowledged theory for wide range of commonsense inferences. [18]

Taxonomic reasoning

Taxonomy is the collection of individuals and categories and their relations. Three basic relations are:

Transitivity is one type of inference in taxonomy. Since Tweety is an instance of robin and robin is a subset of bird, it follows that Tweety is an instance of bird. Inheritance is another type of inference. Since Tweety is an instance of robin, which is a subset of bird and bird is marked with property canfly, it follows that Tweety and robin have property canfly. When an individual taxonomizes more abstract categories, outlining and delimiting specific categories becomes more problematic. Simple taxonomic structures are frequently used in AI programs. For instance, WordNet is a resource including a taxonomy, whose elements are meanings of English words. Web mining systems used to collect commonsense knowledge from Web documents focus on taxonomic relations and specifically in gathering taxonomic relations. [1]

Action and change

The theory of action, events and change is another range of the commonsense reasoning. [19] There are established reasoning methods for domains that satisfy the constraints listed below:

Temporal reasoning

Temporal reasoning is the ability to make presumptions about humans' knowledge of times, durations and time intervals. For example, if an individual knows that Mozart was born after Haydn and died earlier than him, they can use their temporal reasoning knowledge to deduce that Mozart had died younger than Haydn. The inferences involved reduce themselves to solving systems of linear inequalities. [20] To integrate that kind of reasoning with concrete purposes, such as natural language interpretation, is more challenging, because natural language expressions have context dependent interpretation. [21] Simple tasks such as assigning timestamps to procedures cannot be done with total accuracy.

Qualitative reasoning

Qualitative reasoning [22] is the form of commonsense reasoning analyzed with certain success. It is concerned with the direction of change in interrelated quantities. For instance, if the price of a stock goes up, the amount of stocks that are going to be sold will go down. If some ecosystem contains wolves and lambs and the number of wolves decreases, the death rate of the lambs will go down as well. This theory was firstly formulated by Johan de Kleer, who analyzed an object moving on a roller coaster. The theory of qualitative reasoning is applied in many spheres such as physics, biology, engineering, ecology, etc. It serves as the basis for many practical programs, analogical mapping, text understanding.

Challenges in automating commonsense reasoning

As of 2014, there are some commercial systems trying to make the use of commonsense reasoning significant. However, they use statistical information as a proxy for commonsense knowledge, where reasoning is absent. Current programs manipulate individual words, but they don't attempt or offer further understanding. According to Ernest Davis and Gary Marcus, five major obstacles interfere with the producing of a satisfactory "commonsense reasoner". [1]

Compared with humans, as of 2018 existing computer programs perform extremely poorly on modern "commonsense reasoning" benchmark tests such as the Winograd Schema Challenge. [23] The problem of attaining human-level competency at "commonsense knowledge" tasks is considered to probably be "AI complete" (that is, solving it would require the ability to synthesize a human-level intelligence). [24] [25] Some researchers believe that supervised learning data is insufficient to produce an artificial general intelligence capable of commonsense reasoning, and have therefore turned to less-supervised learning techniques. [26]

Approaches and techniques

Commonsense's reasoning study is divided into knowledge-based approaches and approaches that are based on machine learning over and using a large data corpora with limited interactions between these two types of approaches [ citation needed ]. There are also crowdsourcing approaches, attempting to construct a knowledge basis by linking the collective knowledge and the input of non-expert people. Knowledge-based approaches can be separated into approaches based on mathematical logic [ citation needed ].

In knowledge-based approaches, the experts are analyzing the characteristics of the inferences that are required to do reasoning in a specific area or for a certain task. The knowledge-based approaches consist of mathematically grounded approaches, informal knowledge-based approaches and large-scale approaches. The mathematically grounded approaches are purely theoretical and the result is a printed paper instead of a program. The work is limited to the range of the domains and the reasoning techniques that are being reflected on. In informal knowledge-based approaches, theories of reasoning are based on anecdotal data and intuition that are results from empirical behavioral psychology. Informal approaches are common in computer programming. Two other popular techniques for extracting commonsense knowledge from Web documents involve Web mining and Crowd sourcing.

COMET (2019), which uses both the OpenAI GPT language model architecture and existing commonsense knowledge bases such as ConceptNet, claims to generate commonsense inferences at a level approaching human benchmarks. Like many other current efforts, COMET over-relies on surface language patterns and is judged to lack deep human-level understanding of many commonsense concepts. Other language-model approaches include training on visual scenes rather than just text, and training on textual descriptions of scenarios involving commonsense physics. [6] [27]

Related Research Articles

Artificial intelligence (AI) is the intelligence of machines or software, as opposed to the intelligence of living beings, primarily of humans. It is a field of study in computer science that develops and studies intelligent machines. Such machines may be called AIs.

In the field of artificial intelligence, the most difficult problems are informally known as AI-complete or AI-hard, implying that the difficulty of these computational problems, assuming intelligence is computational, is equivalent to that of solving the central artificial intelligence problem—making computers as intelligent as people, or strong AI. To call a problem AI-complete reflects an attitude that it would not be solved by a simple specific algorithm.

<span class="mw-page-title-main">Cyc</span> Artificial intelligence project

Cyc is a long-term artificial intelligence project that aims to assemble a comprehensive ontology and knowledge base that spans the basic concepts and rules about how the world works. Hoping to capture common sense knowledge, Cyc focuses on implicit knowledge that other AI platforms may take for granted. This is contrasted with facts one might find somewhere on the internet or retrieve via a search engine or Wikipedia. Cyc enables semantic reasoners to perform human-like reasoning and be less "brittle" when confronted with novel situations.

Knowledge representation and reasoning is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can use to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language. Knowledge representation incorporates findings from psychology about how humans solve problems, and represent knowledge in order to design formalisms that will make complex systems easier to design and build. Knowledge representation and reasoning also incorporates findings from logic to automate various kinds of reasoning.

Natural-language understanding (NLU) or natural-language interpretation (NLI) is a subset of natural-language processing in artificial intelligence that deals with machine reading comprehension. Natural-language understanding is considered an AI-hard problem.

<span class="mw-page-title-main">Douglas Lenat</span> Computer scientist and AI pioneer

Douglas Bruce Lenat was an American computer scientist and researcher in artificial intelligence who was the founder and CEO of Cycorp, Inc. in Austin, Texas.

<span class="mw-page-title-main">Case-based reasoning</span> Process of solving new problems based on the solutions of similar past problems

In artificial intelligence and philosophy, case-based reasoning (CBR), broadly construed, is the process of solving new problems based on the solutions of similar past problems.

<span class="mw-page-title-main">Symbolic artificial intelligence</span> Methods in artificial intelligence research

In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.

In the history of artificial intelligence, neat and scruffy are two contrasting approaches to artificial intelligence (AI) research. The distinction was made in the 1970s and was a subject of discussion until the mid-1980s.

In artificial intelligence research, commonsense knowledge consists of facts about the everyday world, such as "Lemons are sour", or "Cows say moo", that all humans are expected to know. It is currently an unsolved problem in Artificial General Intelligence. The first AI program to address common sense knowledge was Advice Taker in 1959 by John McCarthy.

<span class="mw-page-title-main">Deborah McGuinness</span>

Deborah Louise McGuinness is an American computer scientist and researcher at Rensselaer Polytechnic Institute (RPI). She is a professor of Computer, Cognitive and Web Sciences, Industrial and Systems Engineering, and an endowed chair in the Tetherless World Constellation, a multidisciplinary research institution within RPI that focuses on the study of theories, methods and applications of the World Wide Web. Her fields of expertise include interdisciplinary data integration, artificial intelligence, specifically in knowledge representation and reasoning, description logics, the semantic web, explanation, and trust.

<span class="mw-page-title-main">History of artificial intelligence</span>

The history of artificial intelligence (AI) began in antiquity, with myths, stories and rumors of artificial beings endowed with intelligence or consciousness by master craftsmen. The seeds of modern AI were planted by philosophers who attempted to describe the process of human thinking as the mechanical manipulation of symbols. This work culminated in the invention of the programmable digital computer in the 1940s, a machine based on the abstract essence of mathematical reasoning. This device and the ideas behind it inspired a handful of scientists to begin seriously discussing the possibility of building an electronic brain.

The philosophy of artificial intelligence is a branch of the philosophy of mind and the philosophy of computer science that explores artificial intelligence and its implications for knowledge and understanding of intelligence, ethics, consciousness, epistemology, and free will. Furthermore, the technology is concerned with the creation of artificial animals or artificial people so the discipline is of considerable interest to philosophers. These factors contributed to the emergence of the philosophy of artificial intelligence.

The following outline is provided as an overview of and topical guide to artificial intelligence:

Frames are an artificial intelligence data structure used to divide knowledge into substructures by representing "stereotyped situations". They were proposed by Marvin Minsky in his 1974 article "A Framework for Representing Knowledge". Frames are the primary data structure used in artificial intelligence frame languages; they are stored as ontologies of sets.

James Frederick Allen is a computational linguist recognized for his contributions to temporal logic, in particular Allen's interval algebra. He is interested in knowledge representation, commonsense reasoning, and natural language understanding, believing that "deep language understanding can only currently be achieved by significant hand-engineering of semantically-rich formalisms coupled with statistical preferences". He is the John H. Dessaurer Professor of Computer Science at the University of Rochester

The following outline is provided as an overview of and topical guide to natural-language processing:

The Winograd schema challenge (WSC) is a test of machine intelligence proposed in 2012 by Hector Levesque, a computer scientist at the University of Toronto. Designed to be an improvement on the Turing test, it is a multiple-choice test that employs questions of a very specific structure: they are instances of what are called Winograd schemas, named after Terry Winograd, professor of computer science at Stanford University.

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

References

  1. 1 2 3 4 5 6 7 Ernest Davis; Gary Marcus (2015). "Commonsense Reasoning and Commonsense Knowledge in Artificial Intelligence". Communications of the ACM. Vol. 58, no. 9. pp. 92–103. doi:10.1145/2701413.
  2. McCarthy, John. "Artificial intelligence, logic and formalizing common sense." Philosophical logic and artificial intelligence. Springer, Dordrecht, 1989. 161-190.
  3. Tandon, Niket; Varde, Aparna S.; de Melo, Gerard (22 February 2018). "Commonsense Knowledge in Machine Intelligence". ACM SIGMOD Record. 46 (4): 49–52. doi:10.1145/3186549.3186562.
  4. Matuszek, Cynthia, et al. "Searching for common sense: Populating cyc from the web." UMBC Computer Science and Electrical Engineering Department Collection (2005).
  5. "How to Teach Artificial Intelligence Some Common Sense". Wired. 13 November 2018. Retrieved 11 February 2021.
  6. 1 2 Pavlus, John (30 April 2020). "Common Sense Comes to Computers". Quanta Magazine. Retrieved 3 May 2020.
  7. Davis, Ernest (25 August 2017). "Logical Formalizations of Commonsense Reasoning: A Survey". Journal of Artificial Intelligence Research. 59: 651–723. doi: 10.1613/jair.5339 .
  8. "Cultivating Common Sense | DiscoverMagazine.com". Discover Magazine. 2017. Archived from the original on 25 March 2018. Retrieved 24 March 2018.
  9. Winograd, Terry (January 1972). "Understanding natural language". Cognitive Psychology. 3 (1): 1–191. doi:10.1016/0010-0285(72)90002-3.
  10. "Don't worry: Autonomous cars aren't coming tomorrow (or next year)". Autoweek. 2016. Archived from the original on 25 March 2018. Retrieved 24 March 2018.
  11. Knight, Will (2017). "Boston may be famous for bad drivers, but it's the testing ground for a smarter self-driving car". MIT Technology Review. Archived from the original on 22 August 2020. Retrieved 27 March 2018.
  12. Prakken, Henry (31 August 2017). "On the problem of making autonomous vehicles conform to traffic law". Artificial Intelligence and Law. 25 (3): 341–363. doi: 10.1007/s10506-017-9210-0 .
  13. Thomason, Richmond (2003-08-27). "Logic and Artificial Intelligence".{{cite journal}}: Cite journal requires |journal= (help)
  14. "Artificial intelligence Programs".
  15. "Artificial intelligence applications".
  16. "Bar Hillel Artificial Intelligence Research Machine Translation". TheGuardian.com .
  17. Antol, Stanislaw, et al. "Vqa: Visual question answering." Proceedings of the IEEE international conference on computer vision. 2015.
  18. "Taxonomy".
  19. "Action and change in Commonsense reasoning".
  20. "Temporal reasoning".
  21. Liu, Hugo, and Push Singh. "Commonsense reasoning in and over natural language Archived 2017-08-09 at the Wayback Machine ." International Conference on Knowledge-Based and Intelligent Information and Engineering Systems. Springer, Berlin, Heidelberg, 2004.
  22. "Qualitative reasoning".
  23. "The Winograd Schema Challenge". cs.nyu.edu. Retrieved 9 January 2018.
  24. Yampolskiy, Roman V. "AI-Complete, AI-Hard, or AI-Easy-Classification of Problems in AI." MAICS. 2012.
  25. Andrich, C, Novosel, L, and Hrnkas, B. (2009). Common Sense Knowledge. Information Search and Retrieval, 2009.
  26. Smith, Craig S. (8 April 2020). "Computers Already Learn From Us. But Can They Teach Themselves?". The New York Times. Retrieved 3 May 2020.
  27. Bosselut, Antoine, et al. "Comet: Commonsense transformers for automatic knowledge graph construction." arXiv preprint arXiv:1906.05317 (2019).