Artificial Intelligence: A Modern Approach

Last updated
Artificial Intelligence: A Modern Approach
Artificial Intelligence- A Modern Approach.jpg
First edition (1995)
Author Stuart J. Russell and Peter Norvig
LanguageEnglish
Genre Computer science
Publisher Prentice Hall
Publication date
2020 (4th Ed.)
Media typebook
Pages1136 (4th Ed.)
ISBN 0-13-461099-7
OCLC 359890490
006.3 20
LC Class Q335 .R86 1995
Website http://aima.cs.berkeley.edu

Artificial Intelligence: A Modern Approach (AIMA) is a university textbook on artificial intelligence, written by Stuart J. Russell and Peter Norvig. It was first published in 1995, and the fourth edition of the book was released on 28 April 2020. [1]

Contents

AIMA has been called "the most popular artificial intelligence textbook in the world", [2] and is considered the standard text in the field of artificial intelligence. [3] [4] As of 2023, it was being used at over 1500 universities worldwide, [5] and it has over 59,000 citations on Google Scholar. [6]

AIMA is intended for an undergraduate audience but can also be used for graduate-level studies with the suggestion of adding some of the primary sources listed in the extensive bibliography.[ citation needed ]

Content

AIMA gives detailed information about the working of algorithms in AI. The book's chapters span from classical AI topics like searching algorithms and first-order logic, propositional logic and probabilistic reasoning to advanced topics such as multi-agent systems, constraint satisfaction problems, optimization problems, artificial neural networks, deep learning, reinforcement learning, and computer vision. [7]

Code

The authors provide a GitHub repository with implementations of various exercises and algorithms from the book in different programming languages. [7] [8] Programs in the book are presented in pseudo code with implementations in Java, Python, Lisp, JavaScript, and Scala available online. [8]

Editions

The first and last editions of AIMA were published in 1995 and 2020, respectively, with four editions published in total (1995, 2003, 2009, 2020). [9]

The following is a list of the US print editions. For other editions, the publishing date and the colors of the cover can vary. [9]

Various editions have been translated from the original English into several languages, including at least Chinese, French, German, Hungarian, Italian, Romanian, Russian, and Serbian. Note that the latest, 4th edition is, however, available only in English and French. [9]

See also

Related Research Articles

Artificial intelligence (AI) is the intelligence of machines or software, as opposed to the intelligence of living beings, primarily of humans. It is a field of study in computer science that develops and studies intelligent machines. Such machines may be called AIs.

<span class="mw-page-title-main">Eliezer Yudkowsky</span> American AI researcher and writer (born 1979)

Eliezer S. Yudkowsky is an American artificial intelligence researcher and writer on decision theory and ethics, best known for popularizing ideas related to friendly artificial intelligence, including the idea that there might not be a "fire alarm" for AI. He is the founder of and a research fellow at the Machine Intelligence Research Institute (MIRI), a private research nonprofit based in Berkeley, California. His work on the prospect of a runaway intelligence explosion influenced philosopher Nick Bostrom's 2014 book Superintelligence: Paths, Dangers, Strategies.

Best-first search is a class of search algorithms, which explores a graph by expanding the most promising node chosen according to a specified rule.

Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn from data and generalize to unseen data, and thus perform tasks without explicit instructions. Recently, generative artificial neural networks have been able to surpass many previous approaches in performance.

<span class="mw-page-title-main">Symbolic artificial intelligence</span> Methods in artificial intelligence research

In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.

Friendly artificial intelligence is hypothetical artificial general intelligence (AGI) that would have a positive (benign) effect on humanity or at least align with human interests or contribute to fostering the improvement of the human species. It is a part of the ethics of artificial intelligence and is closely related to machine ethics. While machine ethics is concerned with how an artificially intelligent agent should behave, friendly artificial intelligence research is focused on how to practically bring about this behavior and ensuring it is adequately constrained.

In the history of artificial intelligence, neat and scruffy are two contrasting approaches to artificial intelligence (AI) research. The distinction was made in the 70s and was a subject of discussion until the middle 80s.

<span class="mw-page-title-main">Peter Norvig</span> American computer scientist (born 1956)

Peter Norvig is an American computer scientist and Distinguished Education Fellow at the Stanford Institute for Human-Centered AI. He previously served as a director of research and search quality at Google. Norvig is the co-author with Stuart J. Russell of the most popular textbook in the field of AI: Artificial Intelligence: A Modern Approach used in more than 1,500 universities in 135 countries.

<span class="mw-page-title-main">Stuart J. Russell</span> British computer scientist and author (born 1962)

Stuart Jonathan Russell is a British computer scientist known for his contributions to artificial intelligence (AI). He is a professor of computer science at the University of California, Berkeley and was from 2008 to 2011 an adjunct professor of neurological surgery at the University of California, San Francisco. He holds the Smith-Zadeh Chair in Engineering at University of California, Berkeley. He founded and leads the Center for Human-Compatible Artificial Intelligence (CHAI) at UC Berkeley. Russell is the co-author with Peter Norvig of the authoritative textbook of the field of AI: Artificial Intelligence: A Modern Approach used in more than 1,500 universities in 135 countries.

An artificial general intelligence (AGI) is a hypothetical type of intelligent agent which, if realized, could learn to accomplish any intellectual task that human beings or animals can perform. Alternatively, AGI has been defined as an autonomous system that surpasses human capabilities in the majority of economically valuable tasks.Creating AGI is a primary goal of some artificial intelligence research and of companies such as OpenAI, DeepMind, and Anthropic. AGI is a common topic in science fiction and futures studies.

<span class="mw-page-title-main">Intelligent agent</span> Software agent which acts autonomously

In artificial intelligence, an intelligent agent (IA) is an agent acting in an intelligent manner; It perceives its environment, takes actions autonomously in order to achieve goals, and may improve its performance with learning or acquiring knowledge. An intelligent agent may be simple or complex: A thermostat or other control system is considered an example of an intelligent agent, as is a human being, as is any system that meets the definition, such as a firm, a state, or a biome.

<span class="mw-page-title-main">History of artificial intelligence</span>

The history of artificial intelligence (AI) began in antiquity, with myths, stories and rumors of artificial beings endowed with intelligence or consciousness by master craftsmen. The seeds of modern AI were planted by philosophers who attempted to describe the process of human thinking as the mechanical manipulation of symbols. This work culminated in the invention of the programmable digital computer in the 1940s, a machine based on the abstract essence of mathematical reasoning. This device and the ideas behind it inspired a handful of scientists to begin seriously discussing the possibility of building an electronic brain.

The philosophy of artificial intelligence is a branch of the philosophy of mind and the philosophy of computer science that explores artificial intelligence and its implications for knowledge and understanding of intelligence, ethics, consciousness, epistemology, and free will. Furthermore, the technology is concerned with the creation of artificial animals or artificial people so the discipline is of considerable interest to philosophers. These factors contributed to the emergence of the philosophy of artificial intelligence.

In the history of artificial intelligence, an AI winter is a period of reduced funding and interest in artificial intelligence research. The field has experienced several hype cycles, followed by disappointment and criticism, followed by funding cuts, followed by renewed interest years or even decades later.

The following outline is provided as an overview of and topical guide to artificial intelligence:

This is a timeline of artificial intelligence, sometimes alternatively called synthetic intelligence.

In the philosophy of artificial intelligence, GOFAI is classical symbolic AI, as opposed to other approaches, such as neural networks, situated robotics, narrow symbolic AI or neuro-symbolic AI. The term was coined by philosopher John Haugeland in his 1985 book Artificial Intelligence: The Very Idea.

In the field of artificial intelligence (AI), AI alignment research aims to steer AI systems towards humans' intended goals, preferences, or ethical principles. An AI system is considered aligned if it advances its intended objectives. A misaligned AI system pursues some objectives, but not the intended ones.

Robert Wilensky was an American computer scientist and emeritus professor at the UC Berkeley School of Information, with his main focus of research in artificial intelligence.

<span class="mw-page-title-main">Thomas Dean (computer scientist)</span> American computer scientist

Thomas L. Dean is an American computer scientist known for his work in robot planning, probabilistic graphical models, and computational neuroscience. He was one of the first to introduce ideas from operations research and control theory to artificial intelligence. In particular, he introduced the idea of the anytime algorithm and was the first to apply the factored Markov decision process to robotics. He has authored several influential textbooks on artificial intelligence.

References

  1. "Artificial Intelligence: A Modern Approach, 4th edition" . Retrieved 2020-05-11.
  2. Gold, Kevin (2011-06-21). "Norvig vs. Chomsky and the Fight for the Future of AI". www.tor.com. Tor Books Blog. Archived from the original on 2011-06-25. Retrieved 2023-12-11.
  3. "Nobel Week Dialogue 2019". NobelPrize.org. Retrieved 2022-01-19.
  4. John, Comex. "Artificial Intelligence" . Retrieved 7 August 2021.
  5. "1549 Schools That Have Adopted AIMA" . Retrieved 2023-12-11.
  6. "Artificial intelligence: a modern approach" . Retrieved 2023-12-11.
  7. 1 2 "Artificial Intelligence: A Modern Approach, 4th US ed". aima.cs.berkeley.edu. 2022-08-22. Retrieved 2023-12-26.
  8. 1 2 "aimacode". GitHub. Retrieved 2023-04-12.
  9. 1 2 3 Russell, Stuart; Norvig, Peter (2021-12-18). "All Editions and Translations of AI: A Modern Approach". aima.cs.berkeley.edu. Retrieved 2023-12-26.

Bibliography