AWS Graviton

Last updated

AWS Graviton is a family of 64-bit ARM-based CPUs designed by the Amazon Web Services (AWS) subsidiary Annapurna Labs. The processor family is distinguished by its lower energy use relative to x86-64, static clock rates, and omission of simultaneous multithreading. It was designed to be tightly integrated with AWS servers and datacenters, and is not sold outside Amazon. [1]

Contents

In 2018, AWS released the first version of Graviton suitable for open-source and non-performance-critical scripting workloads as part of its A1 instance family. [2] The second generation, AWS Graviton2, was announced in December 2019 as the first of its sixth generation instances, with AWS promising 40% improved price/performance over fifth generation Intel and AMD instances [3] and an average of 72% reduction in power consumption. [4] In May 2022, AWS made available Graviton3 processors as part of its seventh generation EC2 instances, offering a further 25% better compute performance over Graviton2. [5]

Origin

The first Annapurna Labs silicon product launched under the AWS umbrella was the AWS Nitro hardware and supporting hypervisor in November 2017. [6] Following on from Nitro, Annapurna began to develop general-purpose CPUs using its expertise.

The benefits AWS anticipated included:

The first Graviton processor reached these goals. Graviton2 now offers better performance compared to X86-64: 35% faster running Redis, [7] 30% faster running Apache Cassandra, [8] and up to 117% higher throughput for MongoDB. [9] In addition to higher performance, Graviton offers 70% lower power consumption [10] and 20% lower price. [11]

Graviton

Graviton
General information
LaunchedNovember 26, 2018;5 years ago (November 26, 2018)
Performance
Max. CPU clock rate 2.3 GHz
Cache
L1 cache 80 KB per core (48 instructions + 32 data)
L2 cache8 MB total
Architecture and classification
Technology node 16 nm
Instruction set AArch64
Instructions AArch64
Extensions
Physical specifications
Cores
History
Successor(s) Graviton2
Support status
Supported

The first Graviton CPU has 16 Cortex A72 cores, with ARMv8-A ISA including Neon, crc, crypto. The vCPUs are physical cores in a single NUMA domain, running at 2.3 GHz. It also includes hardware acceleration for floating-point math, SIMD, plus AES, SHA-1, SHA-256, GCM, and CRC-32 algorithms. [12]

Only the A1 EC2 instance contains the first version of Graviton. [13]

Graviton2

The Graviton2 CPU has 64 Neoverse N1 cores, with ARMv8.2-A ISA including 2×128 bit Neon, LSE, fp16, rcpc, dotprod, crypto. The vCPUs are physical cores in a single NUMA domain, running at 2.5 GHz. [14]

EC2 instances with Graviton2 CPU: M6g, M6gd, C6g, C6gd, C6gn, R6g, R6gd, T4g, X2gd, G5g, Im4gn, Is4gen, I4g. One or more of these instances are available in 28 AWS regions.

Graviton3

The Graviton3 CPU has 64 Neoverse V1 cores, with ARMv8.4-A ISA including 4x128 bit Neon, 2×256 bit SVE, LSE, rng, bf16, int8, crypto. Organized in a single NUMA domain, all vCPUs are physical cores running at 2.6 GHz. [14] Graviton3 has 8 DDR5-4800 memory channels.

Graviton3 provides up to 25% better compute performance, up to 2x higher floating-point performance, up to 2× faster cryptographic workload performance, up to 3× better performance for machine learning workloads including support for bfloat16, and 50% more memory bandwidth compared to AWS Graviton2 processors. Graviton3-based instances use up to 60% less energy for the same performance than comparable EC2 instances. [15]

Graviton3E is a higher power version of Graviton3. [16]

EC2 instances with Graviton3 CPU: C7g, M7g, R7g; with local disk: C7gd, M7gd, R7gd.

EC2 instances with Graviton3E CPU: C7gn, HPC7g.

Graviton4

The Graviton4 CPU has 96 Neoverse V2 cores, with ARMv9.0-A ISA. [17] It has 2 MB of L2 cache per core (192 MB total), and 12 DDR5-5600 memory channels. Graviton4 supports Arm's Branch Target Identification (BTI).

Amazon claims that Graviton4 is up to 40% faster for databases, 30% faster for web applications, and 45% faster for large Java applications than the Graviton3.

EC2 instances with Graviton4 CPU: R8g.

See also

Related Research Articles

<span class="mw-page-title-main">Reduced instruction set computer</span> Processor executing one instruction in minimal clock cycles

In electronics and computer science, a reduced instruction set computer (RISC) is a computer architecture designed to simplify the individual instructions given to the computer to accomplish tasks. Compared to the instructions given to a complex instruction set computer (CISC), a RISC computer might require more instructions in order to accomplish a task because the individual instructions are written in simpler code. The goal is to offset the need to process more instructions by increasing the speed of each instruction, in particular by implementing an instruction pipeline, which may be simpler to achieve given simpler instructions.

ARM is a family of RISC instruction set architectures (ISAs) for computer processors. Arm Ltd. develops the ISAs and licenses them to other companies, who build the physical devices that use the instruction set. It also designs and licenses cores that implement these ISAs.

In computer architecture, 64-bit integers, memory addresses, or other data units are those that are 64 bits wide. Also, 64-bit central processing units (CPU) and arithmetic logic units (ALU) are those that are based on processor registers, address buses, or data buses of that size. A computer that uses such a processor is a 64-bit computer.

<span class="mw-page-title-main">Amazon Elastic Compute Cloud</span> Cloud computing platform

Amazon Elastic Compute Cloud (EC2) is a part of Amazon.com's cloud-computing platform, Amazon Web Services (AWS), that allows users to rent virtual computers on which to run their own computer applications. EC2 encourages scalable deployment of applications by providing a web service through which a user can boot an Amazon Machine Image (AMI) to configure a virtual machine, which Amazon calls an "instance", containing any software desired. A user can create, launch, and terminate server-instances as needed, paying by the second for active servers – hence the term "elastic". EC2 provides users with control over the geographical location of instances that allows for latency optimization and high levels of redundancy. In November 2010, Amazon switched its own retail website platform to EC2 and AWS.

<span class="mw-page-title-main">HiSilicon</span> Chinese fabless semiconductor manufacturing company, fully owned by Huawei

HiSilicon is a Chinese fabless semiconductor company based in Shenzhen, Guangdong province and wholly owned by Huawei. HiSilicon purchases licenses for CPU designs from ARM Holdings, including the ARM Cortex-A9 MPCore, ARM Cortex-M3, ARM Cortex-A7 MPCore, ARM Cortex-A15 MPCore, ARM Cortex-A53, ARM Cortex-A57 and also for their Mali graphics cores. HiSilicon has also purchased licenses from Vivante Corporation for their GC4000 graphics core.

This is a comparison of ARM instruction set architecture application processor cores designed by ARM Holdings and 3rd parties. It does not include ARM Cortex-R, ARM Cortex-M, or legacy ARM cores.

Annapurna Labs is an Israeli microelectronics company. Since January 2015 it has been a wholly-owned subsidiary of Amazon.com. Amazon reportedly acquired the company for its Amazon Web Services division for US$350–370M.

The ARM Cortex-A72 is a central processing unit implementing the ARMv8-A 64-bit instruction set designed by ARM Holdings' Austin design centre. The Cortex-A72 is a 3-way decode out-of-order superscalar pipeline. It is available as SIP core to licensees, and its design makes it suitable for integration with other SIP cores into one die constituting a system on a chip (SoC). The Cortex-A72 was announced in 2015 to serve as the successor of the Cortex-A57, and was designed to use 20% less power or offer 90% greater performance.

Qualcomm Kryo is a series of custom or semi-custom ARM-based CPUs included in the Snapdragon line of SoCs.

This is a timeline of Amazon Web Services, which offers a suite of cloud computing services that make up an on-demand computing platform.

The ARM Cortex-A76 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre. ARM states a 25% and 35% increase in integer and floating point performance, respectively, over a Cortex-A75 of the previous generation.

The ARM Cortex-A77 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre. ARM announced an increase of 23% and 35% in integer and floating point performance, respectively. Memory bandwidth increased 15% relative to the A76.

<span class="mw-page-title-main">Apple A13</span> System on a chip (SoC) designed by Apple Inc.

The Apple A13 Bionic is a 64-bit ARM-based system on a chip (SoC), designed by Apple Inc. It appears in the iPhone 11, 11 Pro/Pro Max, the 9th generation iPad, the iPhone SE and the Studio Display. Apple states that the two high performance cores are 20% faster with 30% lower power consumption than the Apple A12's, and the four high efficiency cores are 20% faster with 30% lower power consumption than the A12's.

<span class="mw-page-title-main">Ampere Computing</span> American fabless semiconductor company

Ampere Computing LLC is an American fabless semiconductor company based in Santa Clara, California that develops processors for servers operating in large scale environments. Ampere also has offices in: Portland, Oregon; Taipei, Taiwan; Raleigh, North Carolina; Bangalore, India; Warsaw, Poland; and Ho Chi Minh City, Vietnam.

The ARM Cortex-A78 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Ltd.'s Austin centre.

The ARM Cortex-X1 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre as part of ARM's Cortex-X Custom (CXC) program.

<span class="mw-page-title-main">Apple A14</span> System on a chip (SoC) designed by Apple Inc.

The Apple A14 Bionic is a 64-bit ARMv8.5-A system on a chip (SoC), designed by Apple Inc. It appears in the fourth generation iPad Air and tenth generation iPad, as well as iPhone 12 Mini, iPhone 12, iPhone 12 Pro, and iPhone 12 Pro Max. Apple states that the central processing unit (CPU) performs up to 40% faster than the A12, while the graphics processing unit (GPU) is up to 30% faster than the A12. It also includes a 16-core neural engine and new machine learning matrix accelerators that perform twice and ten times as fast, respectively.

<span class="mw-page-title-main">Apple A15</span> System on a chip (SoC) designed by Apple Inc.

The Apple A15 Bionic is a 64-bit ARM-based system on a chip (SoC) designed by Apple Inc. It is used in the iPhone 13 and 13 Mini, iPhone 13 Pro and 13 Pro Max, iPad Mini, iPhone SE, iPhone 14 and 14 Plus and Apple TV 4K.

The ARM Cortex-X2 is a central processing unit implementing the ARMv9-A 64-bit instruction set designed by ARM Holdings' Austin design centre as part of ARM's Cortex-X Custom (CXC) program. It forms part of Arm's Total Compute Solutions 2021 (TCS21) along with Arm's Cortex-A710, Cortex-A510, Mali-G710 and CoreLink CI-700/NI-700.

The ARM Neoverse is a group of 64-bit ARM processor cores licensed by Arm Holdings. The cores are intended for datacenter, edge computing, and high-performance computing use. The group consists of ARM Neoverse V-Series, ARM Neoverse N-Series, and ARM Neoverse E-Series.

References

  1. Simonite, Tom (2018-11-27). "New at Amazon: Its Own Chips for Cloud Computing". Wired. ISSN   1059-1028 . Retrieved 2023-08-09.
  2. Sanders, James (29 November 2018). "FAQ: What Arm servers on AWS mean for your cloud and data center strategy". TechRepublic. Retrieved 17 October 2023.
  3. "Announcing New Amazon EC2 M6g, C6g, and R6g Instances Powered by Next-Generation Arm-based AWS Graviton2 Processors". Amazon Web Services. 2019-12-03. Retrieved 2019-12-03.
  4. "NTT DOCOMO and NEC Reduce Power Consumption for 5G SA Core by an Average of 72% using AWS Graviton2, followed by a Successful Onboarding of 5G SA Core on Hybrid Cloud". Amazon Web Services. 2022-09-29. Retrieved 2022-10-11.
  5. "New – Amazon EC2 C7g Instances, Powered by AWS Graviton3 Processors | AWS News Blog". aws.amazon.com. 23 May 2022. Retrieved 17 October 2023.
  6. Liguori, A (2018). "The Nitro Project–Next Generation AWS Infrastructure" (PDF). Hot Chips: A Symposium on High Performance Chips. Institute of Electrical and Electronics Engineers (IEEE). Retrieved 13 October 2023.
  7. "Gain up to 35% performance benefits for deploying Redis on AWS Graviton2". arm. 2021-07-20.
  8. "Increase performance by up to 30% by deploying Apache Cassandra on AWS Graviton2". arm. 2021-08-18.
  9. "MongoDB performance on Arm Neoverse based AWS Graviton2 processors". arm. 2021-06-09.
  10. "NTT DOCOMO and NEC Reduce Power Consumption for 5G SA Core by an Average of 72% using AWS Graviton2, followed by a Successful Onboarding of 5G SA Core on Hybrid Cloud". nec. 2022-11-29.
  11. "20% lower cost and up to 40% higher performance for M6g, C6g, and R6g instances over M5, C5, and R5 instances respectively". amazon. 2022-03-03.
  12. "Amazon's homegrown 2.3GHz 64-bit Graviton processor was very nearly an AMD Arm CPU". theregister. 2018-11-27.
  13. "Amazon EC2 A1 Instances". Amazon Web Services. 2018-11-26. Retrieved 2022-10-11.
  14. 1 2 "Building for Graviton2 and Graviton3". Amazon Web Services. 2022-09-22. Retrieved 2022-10-10.
  15. "Amazon 2021 Letter to Shareholders". AboutAmazon. 2022-04-14. Retrieved 2022-11-16.
  16. "New Amazon EC2 Instance Types In the Works". AWS News Blog. 2022-11-28. Retrieved 2022-11-29.
  17. "Join the preview for new memory-optimized, AWS Graviton4-powered Amazon EC2 instances (R8g)". 2023-11-28. Retrieved 2023-11-28.