Turing (microarchitecture)

Last updated

Turing
Nvidia@12nm@Turing@TU104@GeForce RTX 2080@S TAIWAN 1841A1 PKYN44.000 TU104-400-A1 DSCx3.jpg
LaunchedSeptember 20, 2018;6 years ago (2018-09-20)
Designed by Nvidia
Manufactured by
Fabrication processTSMC 12FFC
Codename(s)TU10x
TU11x
Product Series
Desktop
Professional/workstation
Server/datacenter
Specifications
Compute
  • 28.5 TFLOPS (FP16)
  • 14.2 TFLOPS (FP32)
[1]
L1 cache96 KB (per SM)
L2 cache2 MB to 6 MB
Memory support GDDR6
HBM2
PCIe support PCIe 3.0
Supported Graphics APIs
DirectX DirectX 12 Ultimate (Feature Level 12_2)
Direct3D Direct3D 12.0
Shader Model Shader Model 6.7
OpenCL OpenCL 3.0
OpenGL OpenGL 4.6
CUDA Compute Capability 7.5
Vulkan Vulkan 1.3
Media Engine
Encode codecs
Decode codecs
Color bit-depth
  • 8-bit
  • 10-bit
Encoder(s) supported NVENC
Display outputs
History
Predecessor Pascal
Variant Volta (datacenter/HPC)
Successor Ampere

Turing is the codename for a graphics processing unit (GPU) microarchitecture developed by Nvidia. It is named after the prominent mathematician and computer scientist Alan Turing. The architecture was first introduced in August 2018 at SIGGRAPH 2018 in the workstation-oriented Quadro RTX cards, [2] and one week later at Gamescom in consumer GeForce 20 series graphics cards. [3] Building on the preliminary work of Volta, its HPC-exclusive predecessor, the Turing architecture introduces the first consumer products capable of real-time ray tracing, a longstanding goal of the computer graphics industry. Key elements include dedicated artificial intelligence processors ("Tensor cores") and dedicated ray tracing processors ("RT cores"). Turing leverages DXR, OptiX, and Vulkan for access to ray tracing. In February 2019, Nvidia released the GeForce 16 series GPUs, which utilizes the new Turing design but lacks the RT and Tensor cores.

Contents

Turing is manufactured using TSMC's 12 nm FinFET semiconductor fabrication process. The high-end TU102 GPU includes 18.6 billion transistors fabricated using this process. [1] Turing also uses GDDR6 memory from Samsung Electronics, and previously Micron Technology.

Details

Die shot of the TU104 GPU used in RTX 2080 cards Nvidia@12nm@Turing@TU104@GeForce RTX 2080@S TAIWAN 1841A1 PKYN44.000 TU104-400-A1 DSCx7 poly@5xExt.jpg
Die shot of the TU104 GPU used in RTX 2080 cards
Die shot of the TU106 GPU used in RTX 2060 cards Nvidia@12nm@Turing@TU106@GeForce RTX 2060@S TAIWAN 1844A1 PM4F79.000 TU106-200A-KA-A1 DSCx5 poly@5xExt.jpg
Die shot of the TU106 GPU used in RTX 2060 cards
Die shot of the TU116 GPU used in GTX 1660 cards Nvidia@12nm@Turing@TU116@GeForce GTX 1660@S TAIWAN 1941A1 PPHV26.000 TU116-300-A1 DSCx5 poly@5xExt.jpg
Die shot of the TU116 GPU used in GTX 1660 cards

The Turing microarchitecture combines multiple types of specialized processor core, and enables an implementation of limited real-time ray tracing. [4] This is accelerated by the use of new RT (ray-tracing) cores, which are designed to process quadtrees and spherical hierarchies, and speed up collision tests with individual triangles.

Features in Turing:

The GDDR6 memory is produced by Samsung Electronics for the Quadro RTX series. [6] The RTX 20 series initially launched with Micron memory chips, before switching to Samsung chips by November 2018. [7]

Rasterization

Nvidia reported rasterization (CUDA) performance gains for existing titles of approximately 30–50% over the previous generation. [8] [9]

Ray-tracing

The ray-tracing performed by the RT cores can be used to produce reflections, refractions and shadows, replacing traditional raster techniques such as cube maps and depth maps. Instead of replacing rasterization entirely, however, the information gathered from ray-tracing can be used to augment the shading with information that is much more photo-realistic, especially in regards to off-camera action. Nvidia said the ray-tracing performance increased about 8 times over the previous consumer architecture, Pascal.

Tensor cores

Generation of the final image is further accelerated by the Tensor cores, which are used to fill in the blanks in a partially rendered image, a technique known as de-noising. The Tensor cores perform the result of deep learning to codify how to, for example, increase the resolution of images generated by a specific application or game. In the Tensor cores' primary usage, a problem to be solved is analyzed on a supercomputer, which is taught by example what results are desired, and the supercomputer determines a method to use to achieve those results, which is then done with the consumer's Tensor cores. These methods are delivered via driver updates to consumers. [8] The supercomputer uses a large number of Tensor cores itself.

Turing dies

Comparison of Turing dies
DieTU102 [10] TU104 [11] TU106 [12] TU116 [13] TU117 [14]
Die size754 mm2545 mm2445 mm2284 mm2200 mm2
Transistors18.6B13.6B10.8B6.6B4.7B
Transistor density24.7 MTr/mm225.0 MTr/mm224.3 MTr/mm223.2 MTr/mm223.5 MTr/mm2
Graphics processing
clusters
66332
Streaming
multiprocessors
7248362416
CUDA cores 46083072230415361024
Texture mapping units 2881921449664
Render output units 9664644832
Tensor cores 576384288
RT cores 724836
L1 cache 6.75 MB4.5 MB3.375 MB2.25 MB1.5 MB
96 KB per SM
L2 cache6 MB4 MB4 MB1.5 MB1 MB

Development

Turing's development platform is called RTX. RTX ray-tracing features can be accessed using Microsoft's DXR, OptiX, as well using Vulkan extensions (the last one being also available on Linux drivers). [15] It includes access to AI-accelerated features through NGX. The Mesh Shader, Shading Rate Image functionalities are accessible using DirectX 12, Vulkan and OpenGL extensions on Windows and Linux platforms. [16]

Windows 10 October 2018 update includes the public release of DirectX Raytracing. [17] [18]

Products using Turing

See also

Related Research Articles

<span class="mw-page-title-main">GeForce</span> Brand of GPUs by Nvidia

GeForce is a brand of graphics processing units (GPUs) designed by Nvidia and marketed for the performance market. As of the GeForce 40 series, there have been eighteen iterations of the design. The first GeForce products were discrete GPUs designed for add-on graphics boards, intended for the high-margin PC gaming market, and later diversification of the product line covered all tiers of the PC graphics market, ranging from cost-sensitive GPUs integrated on motherboards, to mainstream add-in retail boards. Most recently, GeForce technology has been introduced into Nvidia's line of embedded application processors, designed for electronic handhelds and mobile handsets.

<span class="mw-page-title-main">Alienware</span> American computer hardware subsidiary of Dell Inc.

Alienware Corporation is an American computer hardware subsidiary brand of Dell. Their product range is dedicated to gaming computers and accessories and can be identified by their alien-themed designs. Alienware was founded in 1996 by Nelson Gonzalez and Alex Aguila. The development of the company is also associated with Frank Azor, Arthur Lewis, Joe Balerdi, and Michael S. Dell (CEO). The company's corporate headquarters is located in The Hammocks, Miami, Florida.

<span class="mw-page-title-main">Quadro</span> Brand of Nvidia graphics cards used in workstations

Quadro was Nvidia's brand for graphics cards intended for use in workstations running professional computer-aided design (CAD), computer-generated imagery (CGI), digital content creation (DCC) applications, scientific calculations and machine learning from 2000 to 2020.

<span class="mw-page-title-main">Ray-tracing hardware</span> Type of 3D graphics accelerator

Ray-tracing hardware is special-purpose computer hardware designed for accelerating ray tracing calculations.

PureVideo is Nvidia's hardware SIP core that performs video decoding. PureVideo is integrated into some of the Nvidia GPUs, and it supports hardware decoding of multiple video codec standards: MPEG-2, VC-1, H.264, HEVC, and AV1. PureVideo occupies a considerable amount of a GPU's die area and should not be confused with Nvidia NVENC. In addition to video decoding on chip, PureVideo offers features such as edge enhancement, noise reduction, deinterlacing, dynamic contrast enhancement and color enhancement.

The GeForce 600 series is a series of graphics processing units developed by Nvidia, first released in 2012. It served as the introduction of the Kepler architecture. It is succeeded by the GeForce 700 series.

The GeForce 700 series is a series of graphics processing units developed by Nvidia. While mainly a refresh of the Kepler microarchitecture, some cards use Fermi (GF) and later cards use Maxwell (GM). GeForce 700 series cards were first released in 2013, starting with the release of the GeForce GTX Titan on February 19, 2013, followed by the GeForce GTX 780 on May 23, 2013. The first mobile GeForce 700 series chips were released in April 2013.

The GeForce 10 series is a series of graphics processing units developed by Nvidia, initially based on the Pascal microarchitecture announced in March 2014. This design series succeeded the GeForce 900 series, and is succeeded by the GeForce 16 series and GeForce 20 series using the Turing microarchitecture.

<span class="mw-page-title-main">Maxwell (microarchitecture)</span> GPU microarchitecture by Nvidia

Maxwell is the codename for a GPU microarchitecture developed by Nvidia as the successor to the Kepler microarchitecture. The Maxwell architecture was introduced in later models of the GeForce 700 series and is also used in the GeForce 800M series, GeForce 900 series, and Quadro Mxxx series, as well as some Jetson products.

<span class="mw-page-title-main">NVLink</span> High speed chip interconnect

NVLink is a wire-based serial multi-lane near-range communications link developed by Nvidia. Unlike PCI Express, a device can consist of multiple NVLinks, and devices use mesh networking to communicate instead of a central hub. The protocol was first announced in March 2014 and uses a proprietary high-speed signaling interconnect (NVHS).

<span class="mw-page-title-main">Pascal (microarchitecture)</span> GPU microarchitecture by Nvidia

Pascal is the codename for a GPU microarchitecture developed by Nvidia, as the successor to the Maxwell architecture. The architecture was first introduced in April 2016 with the release of the Tesla P100 (GP100) on April 5, 2016, and is primarily used in the GeForce 10 series, starting with the GeForce GTX 1080 and GTX 1070, which were released on May 27, 2016, and June 10, 2016, respectively. Pascal was manufactured using TSMC's 16 nm FinFET process, and later Samsung's 14 nm FinFET process.

Graphics Double Data Rate 6 Synchronous Dynamic Random-Access Memory is a type of synchronous graphics random-access memory (SGRAM) with a high bandwidth, "double data rate" interface, designed for use in graphics cards, game consoles, and high-performance computing. It is a type of GDDR SDRAM, and is the successor to GDDR5. Just like GDDR5X it uses QDR in reference to the write command clock (WCK) and ODR in reference to the command clock (CK).

<span class="mw-page-title-main">Nvidia RTX</span> Development platform for rendering graphics

Nvidia RTX is a professional visual computing platform created by Nvidia, primarily used in workstations for designing complex large-scale models in architecture and product design, scientific visualization, energy exploration, and film and video production, as well as being used in mainstream PCs for gaming.

<span class="mw-page-title-main">GeForce 20 series</span> Series of GPUs by Nvidia

The GeForce 20 series is a family of graphics processing units developed by Nvidia. Serving as the successor to the GeForce 10 series, the line started shipping on September 20, 2018, and after several editions, on July 2, 2019, the GeForce RTX Super line of cards was announced.

<span class="mw-page-title-main">GeForce 16 series</span> Series of GPUs by Nvidia

The GeForce 16 series is a series of graphics processing units (GPUs) developed by Nvidia, based on the Turing microarchitecture, announced in February 2019. The 16 series, commercialized within the same timeframe as the 20 series, aims to cover the entry-level to mid-range market, not addressed by the latter. As a result, the media have mainly compared it to AMD's Radeon RX 500 series of GPUs.

Ampere is the codename for a graphics processing unit (GPU) microarchitecture developed by Nvidia as the successor to both the Volta and Turing architectures. It was officially announced on May 14, 2020 and is named after French mathematician and physicist André-Marie Ampère.

<span class="mw-page-title-main">GeForce 30 series</span> GPU series by Nvidia

The GeForce 30 series is a suite of graphics processing units (GPUs) designed and marketed by Nvidia, succeeding the GeForce 20 series. The GeForce 30 series is based on the Ampere architecture, which features Nvidia's second-generation ray tracing (RT) cores and third-generation Tensor Cores. Through Nvidia RTX, hardware-enabled real-time ray tracing is possible on GeForce 30 series cards.

<span class="mw-page-title-main">GeForce 40 series</span> Series of graphics processing units developed by Nvidia

The GeForce 40 series is the most recent family of consumer-level graphics processing units developed by Nvidia, succeeding the GeForce 30 series. The series was announced on September 20, 2022, at the GPU Technology Conference (GTC) 2022 event, and launched on October 12, 2022, starting with its flagship model, the RTX 4090.

Ada Lovelace, also referred to simply as Lovelace, is a graphics processing unit (GPU) microarchitecture developed by Nvidia as the successor to the Ampere architecture, officially announced on September 20, 2022. It is named after the English mathematician Ada Lovelace, one of the first computer programmers. Nvidia announced the architecture along with the GeForce RTX 40 series consumer GPUs and the RTX 6000 Ada Generation workstation graphics card. The Lovelace architecture is fabricated on TSMC's custom 4N process which offers increased efficiency over the previous Samsung 8 nm and TSMC N7 processes used by Nvidia for its previous-generation Ampere architecture.

References

  1. 1 2 "Nvidia Turing GPU Architecture: Graphics Reinvented" (PDF). Nvidia. 2018. Retrieved June 28, 2019.
  2. Smith, Ryan (August 13, 2018). "NVIDIA Reveals Next-Gen Turing GPU Architecture: NVIDIA Doubles-Down on Ray Tracing, GDDR6, & More". AnandTech. Retrieved April 9, 2023.
  3. Smith, Ryan (August 20, 2018). "NVIDIA Announces the GeForce RTX 20 Series: RTX 2080 Ti & 2080 on Sept. 20th, RTX 2070 in October". AnandTech. Retrieved April 9, 2023.
  4. Warren, Tom (August 20, 2018). "Nvidia announces RTX 2000 GPU series with '6 times more performance' and ray-tracing". The Verge. Retrieved August 20, 2018.
  5. Oh, Nate (September 14, 2018). "The NVIDIA Turing GPU Architecture Deep Dive: Prelude to GeForce RTX". AnandTech. Retrieved April 9, 2023.
  6. Mujtaba, Hassan (August 14, 2018). "Samsung GDDR6 Memory Powers NVIDIA's Turing GPU Based Quadro RTX Cards". Wccftech. Retrieved April 9, 2023.
  7. Maislinger, Florian (November 21, 2018). "Faulty RTX 2080 Ti: Nvidia switches from Micron to Samsung for GDDR6 memory". PC Builder's Club. Retrieved July 15, 2019.
  8. 1 2 "#BeForTheGame". Twitch.
  9. Fisher, Jeff (August 20, 2018). "GeForce RTX Propels PC Gaming's Golden Age with Real-Time Ray Tracing". Nvidia. Retrieved April 9, 2023.
  10. "NVIDIA TU102 GPU Specs". TechPowerUp.
  11. "NVIDIA TU104 GPU Specs". TechPowerUp.
  12. "NVIDIA TU106 GPU Specs". TechPowerUp.
  13. "NVIDIA TU116 GPU Specs". TechPowerUp.
  14. "NVIDIA TU117 GPU Specs". TechPowerUp.
  15. "NVIDIA RTX platform". Nvidia. July 20, 2018. Retrieved April 9, 2023.
  16. "Turing Extensions for Vulkan and OpenGL". Nvidia. September 11, 2018. Retrieved April 9, 2023.
  17. Pelletier, Sean (October 2, 2018). "Windows 10 October 2018 Update a Catalyst for Ray-Traced Games". Nvidia. Retrieved April 9, 2023.
  18. van Rhyn, Jacques (October 2, 2018). "DirectX Raytracing and the Windows 10 October 2018 Update". Microsoft. Retrieved April 9, 2023.