AWS Graviton is a family of 64-bit ARM-based CPUs designed by the Amazon Web Services (AWS) subsidiary Annapurna Labs. The processor family is distinguished by its lower energy use relative to x86-64, static clock rates, and lack of simultaneous multithreading. It was designed to be tightly integrated with AWS servers and datacenters, and is not sold outside Amazon. [1]
In 2018, AWS released the first version of Graviton suitable for open-source and non-performance-critical scripting workloads as part of its A1 instance family. [2] The second generation, AWS Graviton2, was announced in December 2019 as the first of its sixth generation instances, with AWS promising 40% improved price/performance over fifth generation Intel and AMD instances [3] and an average of 72% reduction in power consumption. [4] In May 2022, AWS made available Graviton3 processors as part of its seventh generation EC2 instances, offering a further 25% better compute performance over Graviton2. [5]
The first Annapurna Labs silicon product launched under the AWS umbrella was the AWS Nitro hardware and supporting hypervisor in November 2017. [6] Following on from Nitro, Annapurna began to develop general-purpose CPUs using its expertise.
The benefits AWS anticipated included:
The first Graviton processor reached these goals. Graviton2 now offers better performance compared to X86-64: 35% faster running Redis, [7] 30% faster running Apache Cassandra, [8] and up to 117% higher throughput for MongoDB. [9] In addition to higher performance, Graviton offers 70% lower power consumption [10] and 20% lower price. [11]
General information | |
---|---|
Launched | November 26, 2018 |
Performance | |
Max. CPU clock rate | 2.3 GHz |
Cache | |
L1 cache | 80 KB per core (48 instructions + 32 data) |
L2 cache | 8 MB total |
Architecture and classification | |
Technology node | 16 nm |
Instruction set | AArch64 |
Instructions | AArch64 |
Extensions | |
Physical specifications | |
Cores |
|
History | |
Successor | Graviton2 |
Support status | |
Supported |
The first Graviton CPU has 16 Cortex A72 cores, with ARMv8-A ISA including Neon, crc, crypto. The vCPUs are physical cores in a single NUMA domain, running at 2.3 GHz. It also includes hardware acceleration for floating-point math, SIMD, plus AES, SHA-1, SHA-256, GCM, and CRC-32 algorithms. [12]
Only the A1 EC2 instance contains the first version of Graviton. [13]
The Graviton2 CPU has 64 Neoverse N1 cores, with ARMv8.2-A ISA including 2×128 bit Neon, LSE, fp16, rcpc, dotprod, crypto. The vCPUs are physical cores in a single NUMA domain, running at 2.5 GHz. [14]
EC2 instances with Graviton2 CPU: M6g, M6gd, C6g, C6gd, C6gn, R6g, R6gd, T4g, X2gd, G5g, Im4gn, Is4gen, I4g. [15] One or more of these instances are available in 28 AWS regions.
The Graviton3 CPU has 64 Neoverse V1 cores, with ARMv8.4-A ISA including 4×128 bit Neon, 2×256 bit SVE, LSE, rng, bf16, int8, crypto. Organized in a single NUMA domain, all vCPUs are physical cores running at 2.6 GHz. [14] Graviton3 has 8 DDR5-4800 memory channels.
Compared to Graviton2, Graviton3 provides up to 25% better compute performance, up to 2× higher floating-point performance, up to 2× faster cryptographic workload performance, up to 3× better performance for machine learning workloads including support for bfloat16, and 50% more memory bandwidth. Graviton3-based instances use up to 60% less energy for the same performance than comparable EC2 instances. [16]
Graviton3E is a higher power version of Graviton3. [17]
EC2 instances with Graviton3 CPU: C7g, M7g, R7g; with local disk: C7gd, M7gd, R7gd.
EC2 instances with Graviton3E CPU: C7gn, HPC7g.
The Graviton4 CPU has 96 Neoverse V2 cores, with ARMv9.0-A ISA. [18] It has 2 MB of L2 cache per core (192 MB total), and 12 DDR5-5600 memory channels. Graviton4 supports Arm's Branch Target Identification (BTI).
Amazon claims that Graviton4 is up to 40% faster for databases, 30% faster for web applications, and 45% faster for large Java applications than the Graviton3.
EC2 instances with Graviton4 CPU: R8g [19] , X8g [20] , C8g [21] , M8g [22] , I8g [23] .
In electronics and computer science, a reduced instruction set computer (RISC) is a computer architecture designed to simplify the individual instructions given to the computer to accomplish tasks. Compared to the instructions given to a complex instruction set computer (CISC), a RISC computer might require more instructions in order to accomplish a task because the individual instructions are written in simpler code. The goal is to offset the need to process more instructions by increasing the speed of each instruction, in particular by implementing an instruction pipeline, which may be simpler to achieve given simpler instructions.
ARM is a family of RISC instruction set architectures (ISAs) for computer processors. Arm Holdings develops the ISAs and licenses them to other companies, who build the physical devices that use the instruction set. It also designs and licenses cores that implement these ISAs.
Amazon Web Services, Inc. (AWS) is a subsidiary of Amazon that provides on-demand cloud computing platforms and APIs to individuals, companies, and governments, on a metered, pay-as-you-go basis. Clients will often use this in combination with autoscaling. These cloud computing web services provide various services related to networking, compute, storage, middleware, IoT and other processing capacity, as well as software tools via AWS server farms. This frees clients from managing, scaling, and patching hardware and operating systems. One of the foundational services is Amazon Elastic Compute Cloud (EC2), which allows users to have at their disposal a virtual cluster of computers, with extremely high availability, which can be interacted with over the internet via REST APIs, a CLI or the AWS console. AWS's virtual computers emulate most of the attributes of a real computer, including hardware central processing units (CPUs) and graphics processing units (GPUs) for processing; local/RAM memory; hard-disk (HDD)/SSD storage; a choice of operating systems; networking; and pre-loaded application software such as web servers, databases, and customer relationship management (CRM).
Amazon Elastic Compute Cloud (EC2) is a part of Amazon's cloud-computing platform, Amazon Web Services (AWS), that allows users to rent virtual computers on which to run their own computer applications. EC2 encourages scalable deployment of applications by providing a web service through which a user can boot an Amazon Machine Image (AMI) to configure a virtual machine, which Amazon calls an "instance", containing any software desired. A user can create, launch, and terminate server-instances as needed, paying by the second for active servers – hence the term "elastic". EC2 provides users with control over the geographical location of instances that allows for latency optimization and high levels of redundancy. In November 2010, Amazon switched its own retail website platform to EC2 and AWS.
Amazon Elastic Block Store (EBS) provides raw block-level storage that can be attached to Amazon EC2 instances and is used by Amazon Relational Database Service (RDS). It is one of the two block-storage options offered by AWS, with the other being the EC2 Instance Store.
This is a comparison of ARM instruction set architecture application processor cores designed by ARM Holdings and 3rd parties. It does not include ARM Cortex-R, ARM Cortex-M, or legacy ARM cores.
Annapurna Labs is an Israeli microelectronics company. Since January 2015 it has been a wholly owned subsidiary of Amazon.com. Amazon reportedly acquired the company for its Amazon Web Services division for US$350–370M.
The ARM Cortex-A72 is a central processing unit implementing the ARMv8-A 64-bit instruction set designed by ARM Holdings' Austin design centre. The Cortex-A72 is a 3-way decode out-of-order superscalar pipeline. It is available as SIP core to licensees, and its design makes it suitable for integration with other SIP cores into one die constituting a system on a chip (SoC). The Cortex-A72 was announced in 2015 to serve as the successor of the Cortex-A57, and was designed to use 20% less power or offer 90% greater performance.
Qualcomm Kryo is a series of custom or semi-custom ARM-based CPUs included in the Snapdragon line of SoCs.
This is a timeline of Amazon Web Services, which offers a suite of cloud computing services that make up an on-demand computing platform.
The ARM Cortex-A76 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre. ARM states a 25% and 35% increase in integer and floating point performance, respectively, over a Cortex-A75 of the previous generation.
The ARM Cortex-A77 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre. Released in 2019, ARM announced an increase of 23% and 35% in integer and floating point performance over its predecessor, respectively. Memory bandwidth increased 15% relative to the A76.
The Apple A13 Bionic is a 64-bit ARM-based system on a chip (SoC), designed by Apple Inc., part of the Apple silicon series. It appears in the iPhone 11, 11 Pro/Pro Max, the iPad, the iPhone SE and the Studio Display. Apple states that the two high performance cores are 20% faster with 30% lower power consumption than the Apple A12's, and the four high efficiency cores are 20% faster with 30% lower power consumption than the A12's.
Ampere Computing LLC is an American fabless semiconductor company based in Santa Clara, California that develops processors for servers operating in large scale environments. It was founded in 2017 by Renée James.
The ARM Cortex-A78 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Ltd.'s Austin centre.
The ARM Cortex-X1 is a central processing unit implementing the ARMv8.2-A 64-bit instruction set designed by ARM Holdings' Austin design centre as part of ARM's Cortex-X Custom (CXC) program.
The ARM Cortex-A710 is the successor to the ARM Cortex-A78, being the First-Generation Armv9 “big” Cortex CPU. It is the companion to the ARM Cortex-A510 "LITTLE" efficiency core. It was designed by ARM Ltd.'s Austin centre. It is the fourth and last iteration of Arm's Austin core family. It forms part of Arm's Total Compute Solutions 2021 (TCS21) along with Arm's Cortex-X2, Cortex-A510, Mali-G710 and CoreLink CI-700/NI-700.
The ARM Cortex-X2 is a central processing unit implementing the ARMv9-A 64-bit instruction set designed by ARM Holdings' Austin design centre as part of ARM's Cortex-X Custom (CXC) program. It forms part of Arm's Total Compute Solutions 2021 (TCS21) along with Arm's Cortex-A710, Cortex-A510, Mali-G710 and CoreLink CI-700/NI-700.
The ARM Neoverse is a group of 64-bit ARM processor cores licensed by Arm Holdings. The cores are intended for datacenter, edge computing, and high-performance computing use. The group consists of ARM Neoverse V-Series, ARM Neoverse N-Series, and ARM Neoverse E-Series.
The ARM Cortex-A715 is the second generation ARMv9 "big" Cortex CPU. Compared to its predecessor the Cortex-A710 the Cortex-A715 CPU is noted for having a 20% increase in power efficiency, and 5% improvement in performance. The Cortex-A715 shows comparable performance to the previous generation Cortex-X1 CPU. This generation of chips starting with the A715 drops native 32-bit support. It forms part of Arm's Total Compute Solutions 2022 (TCS22) along with Arm's Cortex-X3, Cortex-A510, Arm Immortalis-G715 and CoreLink CI-700/NI-700.