Dynamic game difficulty balancing

Last updated

Dynamic game difficulty balancing (DGDB), also known as dynamic difficulty adjustment (DDA), adaptive difficulty or dynamic game balancing (DGB), is the process of automatically changing parameters, scenarios, and behaviors in a video game in real-time, based on the player's ability, in order to avoid making the player bored (if the game is too easy) or frustrated (if it is too hard). The goal of dynamic difficulty balancing is to keep the user interested from the beginning to the end, providing a good level of challenge.

Contents

Dynamic game elements

Some elements of a game that might be changed via dynamic difficulty balancing include:

Approaches

[A]s players work with a game, their scores should reflect steady improvement. Beginners should be able to make some progress, intermediate people should get intermediate scores, and experienced players should get high scores ... Ideally, the progression is automatic; players start at the beginner's level and the advanced features are brought in as the computer recognizes proficient play.

Chris Crawford, 1982 [1]

Different approaches are found in the literature to address dynamic game difficulty balancing. In all cases, it is necessary to measure, implicitly or explicitly, the difficulty the user is facing at a given moment. This measure can be performed by a heuristic function, which some authors call "challenge function". This function maps a given game state into a value that specifies how easy or difficult the game feels to the user at a specific moment. Examples of heuristics used are:

... or any metric used to calculate a game score. Chris Crawford said "If I were to make a graph of a typical player's score as a function of time spent within the game, that graph should show a curve sloping smoothly and steadily upward. I describe such a game as having a positive monotonic curve". Games without such a curve seem "either too hard or too easy", he said. [1]

Hunicke and Chapman's approach [2] controls the game environment settings in order to make challenges easier or harder. For example, if the game is too hard, the player gets more weapons, recovers life points faster, or faces fewer opponents. Although this approach may be effective, its application can result in implausible situations. A straightforward approach is to combine such "parameters manipulation" to some mechanisms to modify the behavior of the non-player characters (characters controlled by the computer and usually modeled as intelligent agents). This adjustment, however, should be made with moderation, to avoid the 'rubber band' effect. One example of this effect in a racing game would involve the AI driver's vehicles becoming significantly faster when behind the player's vehicle, and significantly slower while in front, as if the two vehicles were connected by a large rubber band.

A traditional implementation of such an agent's intelligence is to use behavior rules, defined during game development. A typical rule in a fighting game would state "punch opponent if he is reachable, chase him otherwise". Extending such an approach to include opponent modeling can be made through Spronck et al.′s dynamic scripting, [3] [4] which assigns to each rule a probability of being picked. Rule weights can be dynamically updated throughout the game, accordingly to the opponent skills, leading to adaptation to the specific user. With a simple mechanism, rules can be picked that generate tactics that are neither too strong nor too weak for the current player.

Andrade et al. [5] divide the DGB problem into two dimensions: competence (learn as well as possible) and performance (act just as well as necessary). This dichotomy between competence and performance is well known and studied in linguistics, as proposed by Noam Chomsky. [6] Their approach faces both dimensions with reinforcement learning (RL). Offline training is used to bootstrap the learning process. This can be done by letting the agent play against itself (selflearning), other pre-programmed agents, or human players. Then, online learning is used to continually adapt this initially built-in intelligence to each specific human opponent, in order to discover the most suitable strategy to play against him or her. Concerning performance, their idea is to find an adequate policy for choosing actions that provide a good game balance, i.e., actions that keep both agent and human player at approximately the same performance level. According to the difficulty the player is facing, the agent chooses actions with high or low expected performance. For a given situation, if the game level is too hard, the agent does not choose the optimal action (provided by the RL framework), but chooses progressively less and less suboptimal actions until its performance is as good as the player's. Similarly, if the game level becomes too easy, it will choose actions whose values are higher, possibly until it reaches the optimal performance.

Demasi and Cruz [7] built intelligent agents employing genetic algorithms techniques to keep alive agents that best fit the user level. Online coevolution is used in order to speed up the learning process. Online coevolution uses pre-defined models (agents with good genetic features) as parents in the genetic operations, so that the evolution is biased by them. These models are constructed by offline training or by hand, when the agent genetic encoding is simple enough.

Other work in the field of DGB is based on the hypothesis that the player-opponent interaction—rather than the audiovisual features, the context or the genre of the game—is the property that contributes the majority of the quality features of entertainment in a computer game. [8] Based on this fundamental assumption, a metric for measuring the real time entertainment value of predator/prey games was introduced, and established as efficient and reliable by validation against human judgment.

Further studies by Yannakakis and Hallam [9] have shown that artificial neural networks (ANN) and fuzzy neural networks can extract a better estimator of player satisfaction than a human-designed one, given appropriate estimators of the challenge and curiosity (intrinsic qualitative factors for engaging gameplay according to Malone) [10] of the game and data on human players' preferences. The approach of constructing user models of the player of a game that can predict the answers to which variants of the game are more or less fun is defined as Entertainment Modeling. The model is usually constructed using machine learning techniques applied to game parameters derived from player-game interaction [11] and/or statistical features of player's physiological signals recorded during play. [12] This basic approach is applicable to a variety of games, both computer [9] and physical.

Caveats

Designing a game that is fair without being predictable is difficult. [13] Andrew Rollings and Ernest Adams cite an example of a game that changed the difficulty of each level based on how the player performed in several preceding levels. Players noticed this and developed a strategy to overcome challenging levels by deliberately playing badly in the levels before the difficult one. The authors stress the importance of covering up the existence of difficulty adaptation so that players are not aware of it. [14]

Uses in video games

An early example of difficulty balancing can be found in Zanac , developed in 1986 by Compile. The game featured a unique adaptive artificial intelligence, in which the game automatically adjusted the difficulty level according to the player's skill level, rate of fire, and the ship's current defensive status/capability. Earlier than this can be found in Midway's 1975 Gun Fight coin-op game. This head-to-head shoot-em-up would aid whichever player had just been shot, by placing a fresh additional object, such as a Cactus plant, on their half of the play-field making it easier for them to hide.

Archon 's computer opponent slowly adapts over time to help players defeat it. [15] Danielle Bunten designed both M.U.L.E. and Global Conquest to dynamically balance gameplay between players. Random events are adjusted so that the player in first place is never lucky and the last-place player is never unlucky. [16]

The first Crash Bandicoot game and its sequels make use of a "Dynamic Difficulty Adjustment" system, slowing down obstacles, giving extra hit points and adding continue points according to the player's number of deaths. According to the game's lead designer Jason Rubin, the goal was to "help weaker players without changing the game for the better players". [17]

The video game Flow was notable for popularizing the application of mental immersion (also called flow) to video games with its 2006 Flash version. The video game design was based on the master's thesis of one of its authors, and was later adapted to PlayStation 3.

SiN Episodes released in 2006 featured a "Personal Challenge System" where the numbers and toughness of enemies faced would vary based on the performance of the player to ensure the level of challenge and pace of progression through the game. The developer, Ritual Entertainment, claimed that players with widely different levels of ability could finish the game within a small range of time of each other. [18]

In 2005, Resident Evil 4 employed a system called the "Difficulty Scale", unknown to most players, as the only mention of it was in the Official Strategy Guide. This system grades the player's performance on a number scale from 1 to 10, and adjusts both enemy behavior/attacks used and enemy damage/resistance based on the player's performance (such as deaths, critical attacks, etc.). The selected difficulty levels lock players at a certain number; for example, on Normal difficulty, one starts at Grade 4, can move down to Grade 2 if doing poorly, or up to Grade 7 if doing well. The grades between difficulties can overlap. [19]

God Hand , a 2006 video game developed by Clover Studio, directed by Resident Evil 4 director Shinji Mikami, and published by Capcom for the PlayStation 2, features a meter during gameplay that regulates enemy intelligence and strength. This meter increases when the player successfully dodges and attacks opponents, and decreases when the player is hit. The meter is divided into four levels, with the hardest level called "Level DIE." The game also has three difficulties, with the easy difficulty only allowing the meter to ascend to level 2, while the hardest difficulty locks the meter to level DIE. This system also offers greater rewards when defeating enemies at higher levels.

The 2008 video game Left 4 Dead uses an artificial intelligence technology dubbed "The AI Director". [20] The AI Director is used to procedurally generate a different experience for the players each time the game is played. It monitors individual players' performance and how well they work together as a group to pace the game, determining the number of zombies that attack the player and the location of boss infected encounters based on information gathered. The Director also determines how quickly players are moving through the level towards each objective; if it detects that players have remained in one place for too long or are not making enough progress, it will summon a horde of common infected to force any players and AI Characters present to move from their current location and combat the new threat. Besides pacing, the Director also controls some video and audio elements of the game to set a mood for a boss encounter or to draw the players' attention to a certain area. [21] Valve calls the way the Director is working "procedural narrative" because instead of having a difficulty level which just ramps up to a constant level, the A.I. analyzes how the players fared in the game so far, and try to add subsequent events that would give them a sense of narrative. [22]

Madden NFL 09 introduces "Madden IQ", which begins with an optional test of the players knowledge of the sport, and abilities in various situations. The score is then used to control the game's difficulty. [23] [24]

In the match-3 game Fishdom , the time limit is adjusted based on how well the player performs. The time limit is increased should the player fail a level, making it possible for any player to beat a level after a few tries.

In the 1999 video game Homeworld , the number of ships that the AI begins with in each mission will be set depending on how powerful the game deems the player's fleet to be. Successful players have larger fleets because they take fewer losses. In this way, a player who is successful over a number of missions will begin to be challenged more and more as the game progresses.

In Fallout: New Vegas and Fallout 3 , as the player increases in level, tougher variants of enemies, enemies with higher statistics and better weapons, or new enemies will replace older ones to retain a constant difficulty, which can be raised, using a slider, with experience bonuses and vice versa in Fallout 3. This can also be done in New Vegas, but there is no bonus to increasing or decreasing the difficulty.

The Mario Kart series features items during races that help an individual driver get ahead of their opponents. These items are distributed based on a driver's position in a way that is an example of dynamic game difficulty balancing. For example, a driver near the bottom of the field is likely to get an item that will drastically increase their speed or sharply decrease the speed of their opponents, whereas a driver in first or second place can expect to get these kinds of items rarely (and will probably receive the game's weaker items). The game's computer racers also adapt to the player's speed - slowing down when the leading player racer is too far behind the best computer racer, and vice versa - as the rival computer racers catch up to the player in first.

Alleged use to shape player buying behaviour

In 2020, a class-action lawsuit in the United States District Court for the Northern District of California accused game developer Electronic Arts of using its patented Dynamic Difficulty Adjustment technology in three of its EA Sports franchises — Madden NFL, FIFA, and NHL — across all games ranging back to the 2017 versions. The plaintiffs say that EA uses this technology to push players into purchasing more loot boxes in the form of Player Packs, saying that it effectively makes even high-stat players not play as well as they should.

The suit also notes that EA uses this technology without disclosing it to players, noting that EA has denied its use in the past in multiple games mentioned in the suit. When asked for comment on the allegations, EA called the claims "baseless" and that they "misrepresent our games." [25] [26] [27] The plaintiffs voluntarily dismissed the lawsuit in 2021. [28]

See also

Related Research Articles

A stealth game is a type of video game in which the player primarily uses stealth to avoid or overcome opponents. Games in the genre typically allow the player to remain undetected by hiding, sneaking, or using disguises. Some games allow the player to choose between a stealthy approach or directly attacking antagonists, but rewarding the player for greater use of stealth. The genre has employed espionage, counter-terrorism, and rogue themes, with protagonists that are special forces operatives, special agents, secret agents, thieves, ninjas, or assassins. Some games have also combined stealth elements with other genres, such as first-person shooters and also platformers.

A game programmer is a software engineer, programmer, or computer scientist who primarily develops codebases for video games or related software, such as game development tools. Game programming has many specialized disciplines, all of which fall under the umbrella term of "game programmer". A game programmer should not be confused with a game designer, who works on game design.

Game balance is a branch of game design with the intention of improving gameplay and user experience by balancing difficulty and fairness. Game balance consists of adjusting rewards, challenges, and/or elements of a game to create the intended player experience.

<i>Populous: The Beginning</i> 1998 video game

Populous: The Beginning is a real-time strategy video game and the third entry in the Populous series, developed by Bullfrog Productions. The game was released in 1998 on Microsoft Windows, and in 1999 for the PlayStation. Unlike earlier games in the series, which cast the player in the role of a god influencing loyal followers, The Beginning took a radical departure and placed the player in the role of a shaman, who directly leads her tribe against opponents. Throughout the twenty-five missions of the campaign, the player leads their tribe across a solar system, dominating enemy tribes and tapping new sources of magic, with the ultimate goal of the shaman attaining godhood herself.

Emergent gameplay refers to complex situations in video games, board games, or role-playing games that emerge from the interaction of relatively simple game mechanics.

In video games, artificial intelligence (AI) is used to generate responsive, adaptive or intelligent behaviors primarily in non-playable characters (NPCs) similar to human-like intelligence. Artificial intelligence has been an integral part of video games since their inception in the 1948, first seen in the game Nim. AI in video games is a distinct subfield and differs from academic AI. It serves to improve the game-player experience rather than machine learning or decision making. During the golden age of arcade video games the idea of AI opponents was largely popularized in the form of graduated difficulty levels, distinct movement patterns, and in-game events dependent on the player's input. Modern games often implement existing techniques such as pathfinding and decision trees to guide the actions of NPCs. AI is often used in mechanisms which are not immediately visible to the user, such as data mining and procedural-content generation. One of the most infamous examples of this NPC technology and gradual difficulty levels can be found in the game Punch-Out!!.

<i>Zanac</i> 1986 video game

Zanac (ザナック) is a shoot 'em up video game developed by Compile and published in Japan by Pony Canyon and in North America by FCI. It was released for the MSX computer, the Family Computer Disk System, the Nintendo Entertainment System, and for the Virtual Console. It was reworked for the MSX2 computer as Zanac EX and for the PlayStation as Zanac X Zanac. Players fly a lone starfighter, dubbed the AFX-6502 Zanac, through twelve levels; their goal is to destroy the System—a part-organic, part-mechanical entity bent on destroying mankind.

In video games, a bot or drone is a type of artificial intelligence (AI)–based expert system software that plays a video game in the place of a human. Bots are used in a variety of video game genres for a variety of tasks: a bot written for a first-person shooter (FPS) works differently from one written for a massively multiplayer online role-playing game (MMORPG). The former may include analysis of the map and even basic strategy; the latter may be used to automate a repetitive and tedious task like farming.

<i>Supremacy: Your Will Be Done</i> 1990 video game

Supremacy: Your Will Be Done, released as Overlord in the United States, is a strategy video game designed by David Perry and Nick Bruty and produced by Probe Software.

<i>NHL 96</i> 1995 ice hockey video game

NHL 96 is a 1995 sports video game developed by EA Tiburon for the SNES, High Score Productions for the Sega Genesis, EA Canada for DOS, and Probe Entertainment for the Game Boy. EA Sports published all versions of the game except the Game Boy version, which was published by THQ. The game is based on the sport of ice hockey and puts the player in control of a hockey team in modes of play such as exhibitions, seasons and playoffs. It is the fifth installment in the NHL game series.

Game Description Language (GDL) is a specialized logic programming language designed by Michael Genesereth. The goal of GDL is to allow the development of AI agents capable of general game playing. It is part of the General Game Playing Project at Stanford University.

General game playing (GGP) is the design of artificial intelligence programs to be able to play more than one game successfully. For many games like chess, computers are programmed to play these games using a specially designed algorithm, which cannot be transferred to another context. For instance, a chess-playing computer program cannot play checkers. General game playing is considered as a necessary milestone on the way to artificial general intelligence.

<i>Magic: The Gathering – Duels of the Planeswalkers</i> 2009 virtual card game

Magic: The Gathering – Duels of the Planeswalkers is a video game based on the popular collectible card game of the same name, published by Wizards of the Coast. It was released on June 17, 2009.

Artificial stupidity is a term used within the field of computer science to refer to a technique of "dumbing down" computer programs in order to deliberately introduce errors in their responses.

Multiplayer online battle arena (MOBA) is a subgenre of strategy video games in which two teams of players compete against each other on a predefined battlefield. Each player controls a single character with a set of distinctive abilities that improve over the course of a game and which contribute to the team's overall strategy. The typical objective is for each team to destroy their opponents' main structure, located at the opposite corner of the battlefield. In some MOBA games, the objective can be defeating every player on the enemy team. Players are assisted by computer-controlled units that periodically spawn in groups and march forward along set paths toward their enemy's base, which is heavily guarded by defensive structures. This type of multiplayer online video games originated as a subgenre of real-time strategy, though MOBA players usually do not construct buildings or units. Moreover, there are examples of MOBA games that are not considered real-time strategy games, such as Smite (2014), and Paragon. The genre is seen as a fusion of real-time strategy, role-playing and action games.

<span class="mw-page-title-main">Robin Hunicke</span> American video game designer and producer (born 1973)

Robin Hunicke is an American video game designer and producer. She is a professor of game design at UC Santa Cruz and the co-founder of Funomena.

<i>Syndicate</i> (2012 video game) 2012 first-person shooter video game

Syndicate is a 2012 first-person shooter game developed by Starbreeze Studios and published by Electronic Arts. The game was released for PlayStation 3, Windows, and Xbox 360 in February 2012.

Since the origin of video games in the early 1970s, the video game industry, the players, and surrounding culture have spawned a wide range of technical and slang terms.

AlphaStar is a computer program by DeepMind that plays the video game StarCraft II. It was unveiled to the public by name in January 2019. In a significant milestone for artificial intelligence, AlphaStar attained Grandmaster status in August 2019.

Artificial intelligence and machine learning techniques are used in video games for a wide variety of applications such as non-player character (NPC) control and procedural content generation (PCG). Machine learning is a subset of artificial intelligence that uses historical data to build predictive and analytical models. This is in sharp contrast to traditional methods of artificial intelligence such as search trees and expert systems.

References

  1. 1 2 Crawford, Chris (December 1982). "Design Techniques and Ideas for Computer Games". BYTE. p. 96. Retrieved 19 October 2013.
  2. Robin Hunicke; V. Chapman (2004). "AI for Dynamic Difficulty Adjustment in Games". Challenges in Game Artificial Intelligence AAAI Workshop. San Jose. pp. 91–96.{{cite book}}: CS1 maint: location missing publisher (link)
  3. Pieter Spronck Archived 2008-12-10 at the Wayback Machine from Tilburg centre for Creative Computing
  4. P. Spronck; I. Sprinkhuizen-Kuyper; E. Postma (2004). "Difficulty Scaling of Game AI". Proceedings of the 5th International Conference on Intelligent Games and Simulation. Belgium. pp. 33–37.{{cite book}}: CS1 maint: location missing publisher (link)
  5. G. Andrade; G. Ramalho; H. Santana; V. Corruble (2005). "Challenge-Sensitive Action Selection: an Application to Game Balancing". Proceedings of the IEEE/WIC/ACM International Conference on Intelligent Agent Technology (IAT-05). Compiègne, France: IEEE Computer Society. pp. 194–200.
  6. Chomsky, Noam. (1965). Aspects of the Theory of Syntax. Cambridge, MA: MIT Press.
  7. P. Demasi; A. Cruz (2002). "Online Coevolution for Action Games". Proceedings of the 3rd International Conference on Intelligent Games and Simulation. London. pp. 113–120.{{cite book}}: CS1 maint: location missing publisher (link)
  8. G. N. Yannakakis; J. Hallam (13–17 July 2004). "Evolving Opponents for Interesting Interactive Computer Games". Proceedings of the 8th International Conference on the Simulation of Adaptive Behavior (SAB'04); From Animals to Animats 8. Los Angeles, California, United States: The MIT Press. pp. 499–508.
  9. 1 2 G. N. Yannakakis; J. Hallam (18–20 May 2006). "Towards Capturing and Enhancing Entertainment in Computer Games". Proceedings of the 4th Hellenic Conference on Artificial Intelligence, Lecture Notes in Artificial Intelligence. Heraklion, Crete, Greece: Springer-Verlag. pp. 432–442.
  10. Malone, T. W. (1981). "What makes computer games fun?". Byte. 6: 258–277.
  11. Wheat, D; Masek, M; Lam, CP; Hingston, P (2015). "Dynamic Difficulty Adjustment in 2D Platformers through Agent-Based Procedural Level Generation". 2015 IEEE International Conference on Systems, Man, and Cybernetics. pp. 2778–2785. doi:10.1109/SMC.2015.485. ISBN   978-1-4799-8697-2. S2CID   19949213.
  12. Chanel, Guillaume; Rebetez, Cyril; Betrancourt, Mireille; Pun, Thierry (2011). "Emotion Assessment from Physiological Signals for Adaptation of Game Difficulty". IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans. 41 (6): 1052–1063. CiteSeerX   10.1.1.650.5420 . doi:10.1109/TSMCA.2011.2116000. S2CID   8681078.
  13. Barry, Tim (1981-05-11). "In Search of the Ultimate Computer Game". InfoWorld. pp. 11, 48. Archived from the original on 2023-02-14. Retrieved 2019-04-17.
  14. A. Rollings; E. Adams. "Gameplay" (PDF). Andrew Rollings and Ernest Adams on Game Design. New Riders Press. Archived from the original (PDF) on 2021-05-01. Retrieved 2014-12-23.
  15. Bateman, Selby (November 1984). "Free Fall Associates: The Designers Behind Archon and Archon II: Adept". Compute!'s Gazette. p. 54. Retrieved 6 July 2014.
  16. "Designing People..." Computer Gaming World. August 1992. pp. 48–54. Archived from the original on 2 July 2014. Retrieved 3 July 2014.
  17. Gavin, Andy (2011-02-07). "Making Crash Bandicoot – part 6". All Things Andy Gavin. Archived from the original on 2011-07-07. Retrieved 2016-09-03.
  18. Monki (2006-05-22). "Monki interviews Tom Mustaine of Ritual about SiN: Emergence". Ain't It Cool News. Archived from the original on 2006-08-23. Retrieved 2006-08-24.
  19. Resident Evil 4: The Official Strategy Guide. Future Press. 4 November 2005.
  20. "Left 4 Dead". Valve. Archived from the original on 2009-03-27.
  21. "Left 4 Dead Hands-on Preview". Left 4 Dead 411. Archived from the original on 2012-02-20. Retrieved 2009-03-16.
  22. Newell, Gabe (21 November 2008). "Gabe Newell Writes for Edge". edge-online.com. Archived from the original on 9 September 2012. Retrieved 2008-11-22.
  23. ""Madden NFL 09 Preseason Report", April 25, 2008". Archived from the original on February 14, 2023. Retrieved May 25, 2015.
  24. ""Madden NFL 09 First Hands On", May 22, 2008". Archived from the original on February 14, 2023. Retrieved May 25, 2015.
  25. Valentine, Rebekah. "EA faces yet another class-action lawsuit connected to loot boxes". GamesIndustry.biz. Archived from the original on 2020-11-12. Retrieved 12 November 2020.
  26. Hetfeld, Malindy (12 November 2020). "Class action lawsuit claims EA's dynamic difficulty tech encourages loot box spending". PC Gamer. Archived from the original on 12 November 2020. Retrieved 12 November 2020.
  27. McAloon, Alissa. "Class action lawsuit accuses EA of changing game difficulty to push loot boxes". www.gamasutra.com. Archived from the original on 11 November 2020. Retrieved 12 November 2020.
  28. Fitzgerald, Jack (2021-02-11). "Notice of Voluntary Dismissal of Action Without Prejudice" (PDF). RECAP Archive.

Further reading