General game playing

Last updated

General game playing (GGP) is the design of artificial intelligence programs to be able to play more than one game successfully. [1] [2] [3] For many games like chess, computers are programmed to play these games using a specially designed algorithm, which cannot be transferred to another context. For instance, a chess-playing computer program cannot play checkers. General game playing is considered as a necessary milestone on the way to artificial general intelligence. [4]

Contents

General video game playing (GVGP) is the concept of GGP adjusted to the purpose of playing video games. For video games, game rules have to be either learnt over multiple iterations by artificial players like TD-Gammon, [5] or are predefined manually in a domain-specific language and sent in advance to artificial players [6] [7] like in traditional GGP. Starting in 2013, significant progress was made following the deep reinforcement learning approach, including the development of programs that can learn to play Atari 2600 games [8] [5] [9] [10] [11] as well as a program that can learn to play Nintendo Entertainment System games. [12] [13] [14]

The first commercial usage of general game playing technology was Zillions of Games in 1998. General game playing was also proposed for trading agents in supply chain management there under price negotiation in online auctions from 2003 on. [15] [16] [17] [18]

History

In 1992, Barney Pell defined the concept of Meta-Game Playing, and developed the "MetaGame" system. This was the first program to automatically generate game rules of chess-like games, and one of the earliest programs to use automated game generation. Pell then developed the system Metagamer. [19] This system was able to play a number of chess-like games, given game rules definition in a special language called Game Description Language (GDL), without any human interaction once the games were generated. [20]

In 1998, the commercial system Zillions of Games was developed by Jeff Mallett and Mark Lefler. The system used a LISP-like language to define the game rules. Zillions of Games derived the evaluation function automatically from the game rules based on piece mobility, board structure and game goals. It also employed usual algorithms as found in computer chess systems: alpha–beta pruning with move ordering, transposition tables, etc. [21] The package was extended in 2007 by the addition of the Axiom plug-in, an alternate metagame engine that incorporates a complete Forth-based programming language.

In 1998, z-Tree was developed by Urs Fischbacher. [22] z-Tree is the first and the most cited software tool for experimental economics. z-Tree allows the definition of game rules in z-Tree-language for game-theoretic experiments with human subjects. It also allows definition of computer players, which participate in a play with human subjects. [23]

In 2005, the Stanford Project General Game Playing was established. [3]

In 2012, the development of PyVGDL started. [24]

GGP implementations

Stanford project

General Game Playing is a project of the Stanford Logic Group of Stanford University, California, which aims to create a platform for general game playing. It is the most well-known effort at standardizing GGP AI, and generally seen as the standard for GGP systems. The games are defined by sets of rules represented in the Game Description Language. In order to play the games, players interact with a game hosting server [25] [26] that monitors moves for legality and keeps players informed of state changes.

Since 2005, there have been annual General Game Playing competitions at the AAAI Conference. The competition judges competitor AI's abilities to play a variety of different games, by recording their performance on each individual game. In the first stage of the competition, entrants are judged on their ability to perform legal moves, gain the upper hand, and complete games faster. In the following runoff round, the AIs face off against each other in increasingly complex games. The AI that wins the most games at this stage wins the competition, and until 2013 its creator used to win a $10,000 prize. [19] So far, the following programs were victorious: [27]

YearNameDeveloperInstitutionRef
2005CluneplayerJim Clune UCLA
2006FluxplayerStephan Schiffel and Michael Thielscher Dresden University of Technology [28]
2007CadiaplayerYngvi Björnsson and Hilmar Finnsson Reykjavik University [29]
2008CadiaplayerYngvi Björnsson, Hilmar Finnsson and Gylfi Þór Guðmundsson Reykjavik University
2009AryJean Méhat Paris 8 University
2010AryJean Méhat Paris 8 University
2011TurboTurtleSam Schreiber
2012CadiaplayerHilmar Finnsson and Yngvi Björnsson Reykjavik University
2013TurboTurtleSam Schreiber
2014SanchoSteve Draper and Andrew Rose [30]
2015GalvaniseRichard Emslie
2016WoodStockEric Piette Artois University

Other approaches

There are other general game playing systems, which use their own languages for defining the game rules. Other general game playing software include:

SystemYearDescription
FRAMASI2009Developed for general game playing and economic experiments during a PhD thesis. [31] [32]
AiAi2015-2017Developed by Stephen Tavener (previous Zillions developer). [33] [34] [35]
PolyGamo Player2017Released by David M. Bennett in September 2017 based on the Unity game engine. [36]
Regular Boardgames2019Developed by Jakub Kowalski, Marek Szykuła, and their team at University of Wrocław. [37] [38]
Ludii2020Released by Cameron Browne and his team at Maastricht University as part of the ERC-funded Digital Ludeme Project. [39] [40] [41]

GVGP implementations

Reinforcement learning

GVGP could potentially be used to create real video game AI automatically, as well as "to test game environments, including those created automatically using procedural content generation and to find potential loopholes in the gameplay that a human player could exploit". [7] GVGP has also been used to generate game rules, and estimate a game's quality based on Relative Algorithm Performance Profiles (RAPP), which compare the skill differentiation that a game allows between good AI and bad AI. [42]

Video Game Description Language

The General Video Game AI Competition (GVGAI) has been running since 2014. In this competition, two-dimensional video games similar to (and sometimes based on) 1980s-era arcade and console games are used instead of the board games used in the GGP competition. It has offered a way for researchers and practitioners to test and compare their best general video game playing algorithms. The competition has an associated software framework including a large number of games written in the Video Game Description Language (VGDL), which should not be confused with GDL and is a coding language using simple semantics and commands that can easily be parsed. One example for VGDL is PyVGDL developed in 2013. [6] [24] The games used in GVGP are, for now, often 2-dimensional arcade games, as they are the simplest and easiest to quantify. [43] To simplify the process of creating an AI that can interpret video games, games for this purpose are written in VGDL manually.[ clarification needed ] VGDL can be used to describe a game specifically for procedural generation of levels, using Answer Set Programming (ASP) and an Evolutionary Algorithm (EA). GVGP can then be used to test the validity of procedural levels, as well as the difficulty or quality of levels based on how an agent performed. [44]

Algorithms

Since GGP AI must be designed to play multiple games, its design cannot rely on algorithms created specifically for certain games. Instead, the AI must be designed using algorithms whose methods can be applied to a wide range of games. The AI must also be an ongoing process, that can adapt to its current state rather than the output of previous states. For this reason, open loop techniques are often most effective. [45]

A popular method for developing GGP AI is the Monte Carlo tree search (MCTS) algorithm. [46] Often used together with the UCT method (Upper Confidence Bound applied to Trees), variations of MCTS have been proposed to better play certain games, as well as to make it compatible with video game playing. [47] [48] [49] Another variation of tree-search algorithms used is the Directed Breadth-first Search (DBS), [50] in which a child node to the current state is created for each available action, and visits each child ordered by highest average reward, until either the game ends or runs out of time. [51] In each tree-search method, the AI simulates potential actions and ranks each based on the average highest reward of each path, in terms of points earned. [46] [51]

Assumptions

In order to interact with games, algorithms must operate under the assumption that games all share common characteristics. In the book Half-Real: Video Games Between Real Worlds and Fictional Worlds, Jesper Juul gives the following definition of games: Games are based on rules, they have variable outcomes, different outcomes give different values, player effort influences outcomes, the player is attached to the outcomes, and the game has negotiable consequences. [52] Using these assumptions, game playing AI can be created by quantifying the player input, the game outcomes, and how the various rules apply, and using algorithms to compute the most favorable path. [43]

See also

Related Research Articles

Artificial intelligence (AI), in its broadest sense, is intelligence exhibited by machines, particularly computer systems. It is a field of research in computer science that develops and studies methods and software that enable machines to perceive their environment and use learning and intelligence to take actions that maximize their chances of achieving defined goals. Such machines may be called AIs.

<span class="mw-page-title-main">Computer Go</span> Field of artificial intelligence around Go computer programs

Computer Go is the field of artificial intelligence (AI) dedicated to creating a computer program that plays the traditional board game Go. The field is sharply divided into two eras. Before 2015, the programs of the era were weak. The best efforts of the 1980s and 1990s produced only AIs that could be defeated by beginners, and AIs of the early 2000s were intermediate level at best. Professionals could defeat these programs even given handicaps of 10+ stones in favor of the AI. Many of the algorithms such as alpha-beta minimax that performed well as AIs for checkers and chess fell apart on Go's 19x19 board, as there were too many branching possibilities to consider. Creation of a human professional quality program with the techniques and hardware of the time was out of reach. Some AI researchers speculated that the problem was unsolvable without creation of human-like AI.

Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn from data and generalize to unseen data and thus perform tasks without explicit instructions. Recently, artificial neural networks have been able to surpass many previous approaches in performance.

<span class="mw-page-title-main">Symbolic artificial intelligence</span> Methods in artificial intelligence research

In artificial intelligence, symbolic artificial intelligence is the term for the collection of all methods in artificial intelligence research that are based on high-level symbolic (human-readable) representations of problems, logic and search. Symbolic AI used tools such as logic programming, production rules, semantic nets and frames, and it developed applications such as knowledge-based systems, symbolic mathematics, automated theorem provers, ontologies, the semantic web, and automated planning and scheduling systems. The Symbolic AI paradigm led to seminal ideas in search, symbolic programming languages, agents, multi-agent systems, the semantic web, and the strengths and limitations of formal knowledge and reasoning systems.

In video games, artificial intelligence (AI) is used to generate responsive, adaptive or intelligent behaviors primarily in non-playable characters (NPCs) similar to human-like intelligence. Artificial intelligence has been an integral part of video games since their inception in the 1950s. AI in video games is a distinct subfield and differs from academic AI. It serves to improve the game-player experience rather than machine learning or decision making. During the golden age of arcade video games the idea of AI opponents was largely popularized in the form of graduated difficulty levels, distinct movement patterns, and in-game events dependent on the player's input. Modern games often implement existing techniques such as pathfinding and decision trees to guide the actions of NPCs. AI is often used in mechanisms which are not immediately visible to the user, such as data mining and procedural-content generation.

John Ross Quinlan is a computer science researcher in data mining and decision theory. He has contributed extensively to the development of decision tree algorithms, including inventing the canonical C4.5 and ID3 algorithms. He also contributed to early ILP literature with First Order Inductive Learner (FOIL). He is currently running the company RuleQuest Research which he founded in 1997.

Game Description Language (GDL) is a specialized logic programming language designed by Michael Genesereth. The goal of GDL is to allow the development of AI agents capable of general game playing. It is part of the General Game Playing Project at Stanford University.

The following outline is provided as an overview of and topical guide to artificial intelligence:

David A. McAllester is an American computer scientist who is Professor and former chief academic officer at the Toyota Technological Institute at Chicago. He received his B.S., M.S. and Ph.D. degrees from the Massachusetts Institute of Technology in 1978, 1979 and 1987 respectively. His PhD was supervised by Gerald Sussman. He was on the faculty of Cornell University for the academic year 1987-1988 and on the faculty of MIT from 1988 to 1995. He was a member of technical staff at AT&T Labs-Research from 1995 to 2002. He has been a fellow of the American Association of Artificial Intelligence since 1997. He has written over 100 refereed publications.

Barney Pell is an American entrepreneur, angel investor and computer scientist. He was co-founder and CEO of Powerset, a pioneering natural language search startup, search strategist and architect for Microsoft's Bing search engine, a pioneer in the field of general game playing in artificial intelligence, and the architect of the first intelligent agent to fly onboard and control a spacecraft. He was co-founder, Vice Chairman and Chief Strategy Officer of Moon Express; co-founder and chairman of LocoMobi; and Associate Founder of Singularity University.

There are a number of competitions and prizes to promote research in artificial intelligence.

<span class="mw-page-title-main">Zillions of Games</span> General game playing software

Zillions of Games is a commercial general game playing system developed by Jeff Mallett and Mark Lefler in 1998. The game rules are specified with S-expressions, Zillions rule language. It was designed to handle mostly abstract strategy board games or puzzles. After parsing the rules of the game, the system's artificial intelligence can automatically play one or more players. It treats puzzles as solitaire games and its AI can be used to solve them.

Machine ethics is a part of the ethics of artificial intelligence concerned with adding or ensuring moral behaviors of man-made machines that use artificial intelligence, otherwise known as artificial intelligent agents. Machine ethics differs from other ethical fields related to engineering and technology. It should not be confused with computer ethics, which focuses on human use of computers. It should also be distinguished from the philosophy of technology, which concerns itself with technology's grander social effects.

In computer science, Monte Carlo tree search (MCTS) is a heuristic search algorithm for some kinds of decision processes, most notably those employed in software that plays board games. In that context MCTS is used to solve the game tree.

<span class="mw-page-title-main">Google DeepMind</span> Artificial intelligence research laboratory

Google DeepMind Technologies Limited is a British-American artificial intelligence research laboratory which serves as a subsidiary of Google. Founded in the UK in 2010, it was acquired by Google in 2014 and merged with Google AI's Google Brain division to become Google DeepMind in April 2023. The company is based in London, with research centres in Canada, France, Germany, and the United States.

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

David Silver is a principal research scientist at Google DeepMind and a professor at University College London. He has led research on reinforcement learning with AlphaGo, AlphaZero and co-lead on AlphaStar.

Deep reinforcement learning is a subfield of machine learning that combines reinforcement learning (RL) and deep learning. RL considers the problem of a computational agent learning to make decisions by trial and error. Deep RL incorporates deep learning into the solution, allowing agents to make decisions from unstructured input data without manual engineering of the state space. Deep RL algorithms are able to take in very large inputs and decide what actions to perform to optimize an objective. Deep reinforcement learning has been used for a diverse set of applications including but not limited to robotics, video games, natural language processing, computer vision, education, transportation, finance and healthcare.

Artificial intelligence and machine learning techniques are used in video games for a wide variety of applications such as non-player character (NPC) control and procedural content generation (PCG). Machine learning is a subset of artificial intelligence that uses historical data to build predictive and analytical models. This is in sharp contrast to traditional methods of artificial intelligence such as search trees and expert systems.

Michael Genesereth is an American logician and computer scientist, who is most known for his work on computational logic and applications of that work in enterprise management, computational law, and general game playing. Genesereth is professor in the Computer Science Department at Stanford University and a professor by courtesy in the Stanford Law School. His 1987 textbook on Logical Foundations of Artificial Intelligence remains one of the key references on symbolic artificial intelligence. He is the author of the influential Game Description Language (GDL) and Knowledge Interchange Format (KIF), the latter of which led to the ISO Common Logic standard.

References

  1. Pell, Barney (1992). H. van den Herik; L. Allis (eds.). "Metagame: a new challenge for games and learning" [Heuristic programming in artificial intelligence 3–the third computerolympiad](PDF). Ellis-Horwood. Archived (PDF) from the original on 2020-02-17. Retrieved 2020-02-17.{{cite journal}}: Cite journal requires |journal= (help)
  2. Pell, Barney (1996). "A Strategic Metagame Player for General Chess-Like Games". Computational Intelligence. 12 (1): 177–198. doi:10.1111/j.1467-8640.1996.tb00258.x. ISSN   1467-8640. S2CID   996006.
  3. 1 2 Genesereth, Michael; Love, Nathaniel; Pell, Barney (15 June 2005). "General Game Playing: Overview of the AAAI Competition". AI Magazine. 26 (2): 62. doi:10.1609/aimag.v26i2.1813. ISSN   2371-9621.
  4. Canaan, Rodrigo; Salge, Christoph; Togelius, Julian; Nealen, Andy (2019). Proceedings of the 14th International Conference on the Foundations of Digital Games[Proceedings of the 14th International Conference on the Leveling the playing field: fairness in AI versus human game benchmarks]. pp. 1–8. doi:10.1145/3337722. ISBN   9781450372176. S2CID   58599284.{{cite book}}: |website= ignored (help)
  5. 1 2 Mnih, Volodymyr; Kavukcuoglu, Koray; Silver, David; Graves, Alex; Antonoglou, Ioannis; Wierstra, Daan; Riedmiller, Martin (2013). "Playing Atari with Deep Reinforcement Learning" (PDF). Neural Information Processing Systems Workshop 2013. Archived (PDF) from the original on 12 September 2014. Retrieved 25 April 2015.
  6. 1 2 Schaul, Tom (August 2013). "A video game description language for model-based or interactive learning". 2013 IEEE Conference on Computational Inteligence in Games (CIG). pp. 1–8. CiteSeerX   10.1.1.360.2263 . doi:10.1109/CIG.2013.6633610. ISBN   978-1-4673-5311-3. S2CID   812565.
  7. 1 2 Levine, John; Congdon, Clare Bates; Ebner, Marc; Kendall, Graham; Lucas, Simon M.; Miikkulainen, Risto; Schaul, Tom; Thompson, Tommy (2013). "General Video Game Playing". Artificial and Computational Intelligence in Games. 6. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik: 77–83. Archived from the original on 9 April 2016. Retrieved 25 April 2015.
  8. Bowling, M.; Veness, J.; Naddaf, Y.; Bellemare, M. G. (2013-06-14). "The Arcade Learning Environment: An Evaluation Platform for General Agents". Journal of Artificial Intelligence Research. 47: 253–279. arXiv: 1207.4708 . doi:10.1613/jair.3912. ISSN   1076-9757. S2CID   1552061.
  9. Mnih, Volodymyr; Kavukcuoglu, Koray; Silver, David; Rusu, Andrei A.; Veness, Joel; Hassabis, Demis; Bellemare, Marc G.; Graves, Alex; Riedmiller, Martin; Fidjeland, Andreas K.; Stig Petersen, Georg Ostrovski; Beattie, Charles; Sadik, Amir; Antonoglou, Ioannis; King, Helen; Kumaran, Dharshan; Wierstra, Daan; Legg, Shane (26 February 2015). "Human-level control through deep reinforcement learning". Nature . 518 (7540): 529–533. Bibcode:2015Natur.518..529M. doi:10.1038/nature14236. PMID   25719670. S2CID   205242740.
  10. Korjus, Kristjan; Kuzovkin, Ilya; Tampuu, Ardi; Pungas, Taivo (2014). "Replicating the Paper "Playing Atari with Deep Reinforcement Learning"" (PDF). University of Tartu. Archived (PDF) from the original on 18 December 2014. Retrieved 25 April 2015.
  11. Guo, Xiaoxiao; Singh, Satinder; Lee, Honglak; Lewis, Richard L.; Wang, Xiaoshi (2014). "Deep Learning for Real-Time Atari Game Play Using Offline Monte-Carlo Tree Search Planning" (PDF). NIPS Proceedingsβ. Conference on Neural Information Processing Systems. Archived (PDF) from the original on 17 November 2015. Retrieved 25 April 2015.
  12. Murphy, Tom (2013). "The First Level of Super Mario Bros. is Easy with Lexicographic Orderings and Time Travel ... after that it gets a little tricky." (PDF). SIGBOVIK. Archived (PDF) from the original on 26 April 2013. Retrieved 25 April 2015.
  13. Murphy, Tom. "learnfun & playfun: A general technique for automating NES games". Archived from the original on 19 April 2015. Retrieved 25 April 2015.
  14. Teller, Swizec (October 28, 2013). "Week 2: Level 1 of Super Mario Bros. is easy with lexicographic orderings and". A geek with a hat. Archived from the original on 30 April 2015. Retrieved 25 April 2015.
  15. McMillen, Colin (2003). "Toward the Development of an Intelligent Agent for the Supply Chain Management Game of the 2003 Trading Agent Competition" [2003 Trading Agent Competition]. Master's Thesis. Minneapolis, MN: University of Minnesota. S2CID   167336006.{{cite journal}}: Cite journal requires |journal= (help)
  16. Zhang, Dongmo (2009). From general game descriptions to a market specification language for general trading agents[Agent-mediated electronic commerce. Designing trading strategies and mechanisms for electronic markets.]. Berlin, Heidelberg: Springer. pp. 259–274. Bibcode:2010aecd.book..259T. CiteSeerX   10.1.1.467.4629 .
  17. "AGAPE - An Auction LanGuage for GenerAl Auction PlayErs". AGAPE (in French). 8 March 2019. Archived from the original on 2 August 2021. Retrieved 5 March 2020.
  18. Michael, Friedrich; Ignatov, Dmitry (2019). "General Game Playing B-to-B Price Negotiations" (PDF). CEUR Workshop Proceedings. -2479: 89–99. Archived (PDF) from the original on 6 December 2019. Retrieved 5 March 2020.
  19. 1 2 Barney Pell's research on computer game playing Archived 2007-08-12 at the Wayback Machine .
  20. "Metagame and General Game Playing". Metagame and General Game Playing. Archived from the original on 3 March 2001. Retrieved 27 March 2016.
  21. Available: Universal Game Engine Archived 2012-11-03 at the Wayback Machine email to comp.ai.games by Jeff Mallett, 10-Dec-1998.
  22. "UZH - z-Tree - Zurich Toolbox for Readymade Economic Experiments". www.ztree.uzh.ch. Archived from the original on 21 February 2016. Retrieved 17 February 2020.
  23. Beckenkamp, Martin; Hennig-Schmidt, Heike; Maier-Rigaud, Frank P. (1 March 2007). "Cooperation in Symmetric and Asymmetric Prisoner's Dilemma Games". Social Science Research Network. SSRN   968942.{{cite journal}}: Cite journal requires |journal= (help)
  24. 1 2 Schaul, Tom (7 February 2020). "schaul/py-vgdl". GitHub . Archived from the original on 11 June 2018. Retrieved 9 February 2020.
  25. GGP Server Archived 2014-02-21 at the Wayback Machine , platform for competition of general game playing systems.
  26. Dresden GGP Server Archived 2013-04-07 at the Wayback Machine , platform for competition of general game playing systems with automatic scheduling of matches.
  27. "General Game Playing". www.general-game-playing.de. Archived from the original on 2008-12-26. Retrieved 2008-08-21.
  28. Information about Fluxplayer Archived 2011-07-19 at the Wayback Machine , the winner of the 2nd International General Game Playing competition.
  29. Information about CADIAPlayer Archived 2011-07-22 at the Wayback Machine , more information about the winner of the 3rd, 4th, and 8th International General Game Playing competitions.
  30. Sancho is GGP Champion 2014! Archived 2015-12-22 at the Wayback Machine , winner of the 2014 International General Game Playing competition.
  31. Tagiew, Rustam (2009). Filipe, Joaquim; Fred, Ana; Sharp, Bernadette (eds.). Towards a framework for management of strategic interaction [Proceedings of the International Conference on Agents and Artificial Intelligence](PDF). Porto, Portugal. pp. 587–590. ISBN   978-989-8111-66-1. Archived (PDF) from the original on 2021-03-09. Retrieved 2021-06-02.{{cite book}}: CS1 maint: location missing publisher (link)
  32. Tagiew, Rustam (2011). Strategische Interaktion realer Agenten Ganzheitliche Konzeptualisierung und Softwarekomponenten einer interdisziplinären Forschungsinfrastruktur (neue Ausg ed.). Saarbrücken. ISBN   9783838125121.{{cite book}}: CS1 maint: location missing publisher (link)
  33. "Zillions of Games - Who Are We?". www.zillions-of-games.com. Archived from the original on 2017-11-15. Retrieved 2017-11-16.
  34. "AiAi Home Page – Stephen Tavener". mrraow.com. Archived from the original on 2015-09-06. Retrieved 2017-11-16.
  35. "Ai Ai announcement thread". BoardGameGeek. Archived from the original on 2017-11-16. Retrieved 2017-11-16.
  36. "The PolyGamo Player Project | Programming Languages and General Players for Abstract Games and Puzzles". www.polyomino.com. Archived from the original on 2002-09-23. Retrieved 2017-11-16.
  37. Kowalski, Jakub; Mika, Maksymilian; Sutowicz, Jakub; Szykuła, Marek (2019-07-17). "Regular Boardgames". Proceedings of the AAAI Conference on Artificial Intelligence. 33 (1): 1699–1706. doi: 10.1609/aaai.v33i01.33011699 . ISSN   2374-3468. S2CID   20296467.
  38. Kowalski, Jakub; Miernik, Radoslaw; Mika, Maksymilian; Pawlik, Wojciech; Sutowicz, Jakub; Szykula, Marek; Tkaczyk, Andrzej (2020). "Efficient Reasoning in Regular Boardgames". 2020 IEEE Conference on Games (CoG). pp. 455–462. arXiv: 2006.08295 . doi:10.1109/cog47356.2020.9231668. ISBN   978-1-7281-4533-4. S2CID   219687404 . Retrieved 2023-11-19.
  39. "Ludii Portal | Home of the Ludii General Game System". www.ludii.games. Archived from the original on 2021-10-27. Retrieved 2021-10-27.
  40. "Digital Ludeme Project | Modelling the Evolution of Traditional Games". www.ludeme.eu. Archived from the original on 2021-10-02. Retrieved 2021-10-27.
  41. Piette, E.; Soemers, D. J. N. J.; Stephenson, M.; Sironi, C.; Stephenson, M.; Winands M. H. M.; Browne, C. (2020). "Ludii – The Ludemic General Game System" (PDF). European Conference on Artificial Intelligence (ECAI 2020), Santiago de Compestela. Archived (PDF) from the original on 2022-01-21. Retrieved 2021-10-27.
  42. Nielsen, Thorbjørn S.; Barros, Gabriella A. B.; Togelius, Julian; Nelson, Mark J. "Towards generating arcade game rules with VGDL" (PDF). Archived (PDF) from the original on 2015-09-12. Retrieved 2018-02-24.
  43. 1 2 Levine, John; Congdon, Clare Bates; Ebner, Marc; Kendall, Graham; Lucas, Simon M.; Miikkulainen Risto, Schaul; Tom, Thompson; Tommy. "General Video Game Playing" (PDF). Archived (PDF) from the original on 2016-04-18. Retrieved 2016-04-09.
  44. Neufeld, Xenija; Mostaghim, Sanaz; Perez-Liebana, Diego. "Procedural Level Generation with Answer Set Programming for General Video Game Playing" (PDF). Archived (PDF) from the original on 2016-03-28. Retrieved 2018-02-24.
  45. Świechowski, Maciej; Park, Hyunsoo; Mańdziuk, Jacek; Kim, Kyung-Joong (2015). "Recent Advances in General Game Playing". The Scientific World Journal. 2015. Hindawi Publishing Corporation: 986262. doi: 10.1155/2015/986262 . PMC   4561326 . PMID   26380375.
  46. 1 2 "Monte-Carlo Tree Search for General Game Playing". ResearchGate. Retrieved 2016-04-01.
  47. Finnsson, Hilmar (2012). "Generalized Monte-Carlo Tree Search Extensions for General Game Playing". Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence. Archived from the original on 2013-10-15. Retrieved 2016-04-09.
  48. Frydenberg, Frederik; Anderson, Kasper R.; Risi, Sebastian; Togelius, Julian. "Investigating MCTS Modifications in General Video Game Playing" (PDF). Archived (PDF) from the original on 2016-04-12. Retrieved 2016-04-09.
  49. M. Swiechowski; J. Mandziuk; Y. S. Ong, "Specialization of a UCT-based General Game Playing Program to Single-Player Games," in IEEE Transactions on Computational Intelligence and AI in Games, vol.PP, no.99, pp.1-1 doi : 10.1109/TCIAIG.2015.2391232
  50. "Changing the root node from a previous game step". Archived from the original on 2021-01-17. DBS: A Directed Breadth First Search (DBS) algorithm
  51. 1 2 Perez, Diego; Dieskau, Jens; Hünermund, Martin. "Open Loop Search for General Video Game Playing" (PDF). Archived (PDF) from the original on 2016-03-28. Retrieved 2016-04-09.
  52. Jesper Juul. Half-Real: Video Games Between Real Rules and Fictional Worlds. MIT Press, 2005.