This article has multiple issues. Please help improve it or discuss these issues on the talk page . (Learn how and when to remove these template messages)
|
Pre-rendering is the process in which video footage is not rendered in real-time by the hardware that is outputting or playing back the video. Instead, the video is a recording of footage that was previously rendered on different equipment (typically one that is more powerful than the hardware used for playback). Pre-rendered assets (typically movies) may also be outsourced by the developer to an outside production company. Such assets usually have a level of complexity that is too great for the target platform to render in real-time.
The term pre-rendered refers to anything that is not rendered in real-time. This includes content that could have been run in real-time with more effort on the part of the developer (e.g. video that covers many of a game's environments without pausing to load, or video of a game in an early state of development that is rendered in slow-motion and then played back at regular speed). This term is generally not used to refer to video captures of real-time rendered graphics despite the fact that video is technically pre-rendered by its nature. The term is also not used to refer to hand drawn assets or photographed assets (these assets not being computer rendered in the first place).
The advantage of pre-rendering is the ability to use graphic models that are more complex and computationally intensive than those that can be rendered in real-time, due to the possibility of using multiple computers over extended periods of time to render the end results. For instance, a comparison could be drawn between rail-shooters Maximum Force (which used pre-rendered 3D levels but 2D sprites for enemies) and Virtua Cop (using 3D polygons); Maximum Force was more realistic looking due to the limitations of Virtua Cop's 3D engine, but Virtua Cop has actual depth (able to portray enemies close and far away, along with body-specific hits and multiple hits) compared to the limits of the 2D sprite enemies in Maximum Force. [1]
The disadvantage of pre-rendering, in the case of video game graphics, is a generally lower level of interactivity, if any, with the player. Another negative side of pre-rendered assets is that changes cannot be made during gameplay. A game with pre-rendered backgrounds is forced to use fixed camera angles, and a game with pre-rendered video generally cannot reflect any changes the game's characters might have undergone during gameplay (such as wounds or customized clothing) without having an alternate version of the video stored. This is generally not feasible due to the large amount of space required to store pre-rendered assets of high quality. However, in some advanced implementations, such as in Final Fantasy VIII , real-time assets were composited with pre-rendered video, allowing dynamic backgrounds and changing camera angles. Another problem is that a game with pre-rendered lighting cannot easily change the state of the lighting in a convincing manner.
As the technology continued to advance in the mid-2000s, video game graphics were able to achieve the photorealism that was previously limited to pre-rendering, as seen in the growth of Machinima .
Pre-rendered graphics are used primarily as cutscenes in modern video games, where they are also known as full motion video. The use of pre-rendered 3D computer graphics for video sequences date back to two arcade laserdisc video games introduced in late 1983: Interstellar, [2] [3] introduced by Funai at the AM Show in September, [4] and Star Rider , [5] introduced by Williams Electronics at the AMOA show in October. [6]
The Sharp X68000 enhanced remake of Ys I: Ancient Ys Vanished , released in 1991, used 3D pre-rendered graphics for the boss sprites, though this ended up creating what is considered "a bizarre contrast" with the game's mostly 2D graphics. [7] One of the first games to extensively use pre-rendered graphics along with full motion video was The 7th Guest . Released in 1993 as one of the first PC games exclusively on CD-ROM, the game was hugely popular, although reviews from critics were mixed. The game featured pre-rendered video sequences that were at a resolution of 640x320 at 15 frames per second, a feat previously thought impossible on personal computers. Shortly after, the release of Myst in 1993 made the use of pre-rendered graphics and CD-ROMs even more popular; most of the rendered work of Myst became the basis for the re-make realMyst: Interactive 3D Edition with its free-roaming real-time 3D graphics. The most graphically advanced use of entirely pre-rendered graphics in games is often claimed to be Myst IV: Revelation , released in 2004.
One of the first significant console games with pre-rendered graphics was Donkey Kong Country , released on the SNES in 1994.
The use of pre-rendered backgrounds and movies also was made popular by the Resident Evil and Final Fantasy franchises on the original PlayStation, both of which use pre-rendered backgrounds and movies extensively to provide a visual presentation that is far greater than the console can provide with real-time 3D. These games include real-time elements (characters, items, etc.) in addition to pre-rendered backgrounds to provide interactivity. Often, a game using pre-rendered backgrounds can devote additional processing power to the remaining interactive elements, resulting in a level of detail greater than the norm for the host platform. In some cases, the visual quality of the interactive elements is still far behind the pre-rendered backgrounds.
In the late 1990s and early 2000s, when most 3D game engines had pre-calculated/fixed Lightmaps and texture mapping, developers often turned to pre-rendered graphics which had a much higher level of realism. However this has lost favor since the mid-2000s, as advances in consumer PC and video game console graphics have enabled the use of the game's own engine to render these cinematics. For instance, the id Tech 4 engine used in Doom 3 allowed bump mapping and dynamic per-pixel lighting, previously only found in pre-rendered videos.
Games such as Warcraft III: Reign of Chaos have used both types of cutscenes; pre-rendered for the beginning and end of a campaign, and the in-game engine for level briefings and character dialogue during a mission.
Some games also use 16-bit pre-rendered skybox, like Half-Life (only GoldSrc version), Re-Volt , Quake II , and others.
CG movies such as Toy Story , Shrek and Final Fantasy: The Spirits Within are entirely pre-rendered.
Another increasingly common pre-rendering method is the generation of texture sets for 3D games, which are often used with complex real-time algorithms to simulate extraordinarily high levels of detail. While making Doom 3 , id Software used pre-rendered models as the basis for generating normal, specular and diffuse lighting maps that simulate the detail of the original model in real-time.
Pre-rendered lighting is a technique that is losing popularity. Processor-intensive ray tracing algorithms can be used during a game's production to generate light textures, which are simply applied on top of the usual hand drawn textures.
Rendering or image synthesis is the process of generating a photorealistic or non-photorealistic image from a 2D or 3D model by means of a computer program. The resulting image is referred to as a rendering. Multiple models can be defined in a scene file containing objects in a strictly defined language or data structure. The scene file contains geometry, viewpoint, textures, lighting, and shading information describing the virtual scene. The data contained in the scene file is then passed to a rendering program to be processed and output to a digital image or raster graphics image file. The term "rendering" is analogous to the concept of an artist's impression of a scene. The term "rendering" is also used to describe the process of calculating effects in a video editing program to produce the final video output.
2.5D perspective refers to gameplay or movement in a video game or virtual reality environment that is restricted to a two-dimensional (2D) plane with little or no access to a third dimension in a space that otherwise appears to be three-dimensional and is often simulated and rendered in a 3D digital environment.
In computer graphics, a shader is a computer program that calculates the appropriate levels of light, darkness, and color during the rendering of a 3D scene—a process known as shading. Shaders have evolved to perform a variety of specialized functions in computer graphics special effects and video post-processing, as well as general-purpose computing on graphics processing units.
A first-person shooter engine is a video game engine specialized for simulating 3D environments for use in a first-person shooter video game. First-person refers to the view where the players see the world from the eyes of their characters. Shooter refers to games which revolve primarily around wielding firearms and killing other entities in the game world, either non-player characters or other players.
Real-time computer graphics or real-time rendering is the sub-field of computer graphics focused on producing and analyzing images in real time. The term can refer to anything from rendering an application's graphical user interface (GUI) to real-time image analysis, but is most often used in reference to interactive 3D computer graphics, typically using a graphics processing unit (GPU). One example of this concept is a video game that rapidly renders changing 3D environments to produce an illusion of motion.
Low poly is a polygon mesh in 3D computer graphics that has a relatively small number of polygons. Low poly meshes occur in real-time applications as contrast with high-poly meshes in animated movies and special effects of the same era. The term low poly is used in both a technical and a descriptive sense; the number of polygons in a mesh is an important factor to optimize for performance but can give an undesirable appearance to the resulting graphics.
Software rendering is the process of generating an image from a model by means of computer software. In the context of computer graphics rendering, software rendering refers to a rendering process that is not dependent upon graphics hardware ASICs, such as a graphics card. The rendering takes place entirely in the CPU. Rendering everything with the (general-purpose) CPU has the main advantage that it is not restricted to the (limited) capabilities of graphics hardware, but the disadvantage is that more transistors are needed to obtain the same speed.
Autodesk Softimage, or simply Softimage was a 3D computer graphics application, for producing 3D computer graphics, 3D modeling, and computer animation. Now owned by Autodesk and formerly titled Softimage|XSI, the software has been predominantly used in the film, video game, and advertising industries for creating computer generated characters, objects, and environments.
High-dynamic-range rendering, also known as high-dynamic-range lighting, is the rendering of computer graphics scenes by using lighting calculations done in high dynamic range (HDR). This allows preservation of details that may be lost due to limiting contrast ratios. Video games and computer-generated movies and special effects benefit from this as it creates more realistic scenes than with more simplistic lighting models.
In computer graphics, per-pixel lighting refers to any technique for lighting an image or scene that calculates illumination for each pixel on a rendered image. This is in contrast to other popular methods of lighting such as vertex lighting, which calculates illumination at each vertex of a 3D model and then interpolates the resulting values over the model's faces to calculate the final per-pixel color values.
3D rendering is the 3D computer graphics process of converting 3D models into 2D images on a computer. 3D renders may include photorealistic effects or non-photorealistic styles.
3D computer graphics, sometimes called CGI, 3-D-CGI or three-dimensional computer graphics, are graphics that use a three-dimensional representation of geometric data that is stored in the computer for the purposes of performing calculations and rendering digital images, usually 2D images but sometimes 3D images. The resulting images may be stored for viewing later or displayed in real time.
Computer graphics deals with generating images and art with the aid of computers. Computer graphics is a core technology in digital photography, film, video games, digital art, cell phone and computer displays, and many specialized applications. A great deal of specialized hardware and software has been developed, with the displays of most devices being driven by computer graphics hardware. It is a vast and recently developed area of computer science. The phrase was coined in 1960 by computer graphics researchers Verne Hudson and William Fetter of Boeing. It is often abbreviated as CG, or typically in the context of film as computer generated imagery (CGI). The non-artistic aspects of computer graphics are the subject of computer science research.
Isometric video game graphics are graphics employed in video games and pixel art that use a parallel projection, but which angle the viewpoint to reveal facets of the environment that would otherwise not be visible from a top-down perspective or side view, thereby producing a three-dimensional (3D) effect. Despite the name, isometric computer graphics are not necessarily truly isometric—i.e., the x, y, and z axes are not necessarily oriented 120° to each other. Instead, a variety of angles are used, with dimetric projection and a 2:1 pixel ratio being the most common. The terms "3/4 perspective", "3/4 view", "2.5D", and "pseudo 3D" are also sometimes used, although these terms can bear slightly different meanings in other contexts.
A variety of computer graphic techniques have been used to display video game content throughout the history of video games. The predominance of individual techniques have evolved over time, primarily due to hardware advances and restrictions such as the processing power of central or graphics processing units.
In 3D computer graphics, 3D modeling is the process of developing a mathematical coordinate-based representation of a surface of an object in three dimensions via specialized software by manipulating edges, vertices, and polygons in a simulated 3D space.
The PlayStation technical specifications describe the various components of the original PlayStation video game console.
Volumetric capture or volumetric video is a technique that captures a three-dimensional space, such as a location or performance. This type of volumography acquires data that can be viewed on flat screens as well as using 3D displays and VR goggles. Consumer-facing formats are numerous and the required motion capture techniques lean on computer graphics, photogrammetry, and other computation-based methods. The viewer generally experiences the result in a real-time engine and has direct input in exploring the generated volume.
A cutscene or event scene is a sequence in a video game that is not interactive, interrupting the gameplay. Such scenes are used to show conversations between characters, set the mood, reward the player, introduce newer models and gameplay elements, show the effects of a player's actions, create emotional connections, improve pacing or foreshadow future events.