Advanced Vector Extensions

Last updated

Advanced Vector Extensions (AVX, also known as Gesher New Instructions and then Sandy Bridge New Instructions) are SIMD extensions to the x86 instruction set architecture for microprocessors from Intel and Advanced Micro Devices (AMD). They were proposed by Intel in March 2008 and first supported by Intel with the Sandy Bridge [1] microarchitecture shipping in Q1 2011 and later by AMD with the Bulldozer [2] microarchitecture shipping in Q4 2011. AVX provides new features, new instructions, and a new coding scheme.

Contents

AVX2 (also known as Haswell New Instructions) expands most integer commands to 256 bits and introduces new instructions. They were first supported by Intel with the Haswell microarchitecture, which shipped in 2013.

AVX-512 expands AVX to 512-bit support using a new EVEX prefix encoding proposed by Intel in July 2013 and first supported by Intel with the Knights Landing co-processor, which shipped in 2016. [3] [4] In conventional processors, AVX-512 was introduced with Skylake server and HEDT processors in 2017.

Advanced Vector Extensions

AVX uses sixteen YMM registers to perform a single instruction on multiple pieces of data (see SIMD). Each YMM register can hold and do simultaneous operations (math) on:

The width of the SIMD registers is increased from 128 bits to 256 bits, and renamed from XMM0–XMM7 to YMM0–YMM7 (in x86-64 mode, from XMM0–XMM15 to YMM0–YMM15). The legacy SSE instructions can be still utilized via the VEX prefix to operate on the lower 128 bits of the YMM registers.

AVX-512 register scheme as extension from the AVX (YMM0-YMM15) and SSE (XMM0-XMM15) registers
5112562551281270
  ZMM0    YMM0    XMM0  
ZMM1YMM1XMM1
ZMM2YMM2XMM2
ZMM3YMM3XMM3
ZMM4YMM4XMM4
ZMM5YMM5XMM5
ZMM6YMM6XMM6
ZMM7YMM7XMM7
ZMM8YMM8XMM8
ZMM9YMM9XMM9
ZMM10YMM10XMM10
ZMM11YMM11XMM11
ZMM12YMM12XMM12
ZMM13YMM13XMM13
ZMM14YMM14XMM14
ZMM15YMM15XMM15
ZMM16YMM16XMM16
ZMM17YMM17XMM17
ZMM18YMM18XMM18
ZMM19YMM19XMM19
ZMM20YMM20XMM20
ZMM21YMM21XMM21
ZMM22YMM22XMM22
ZMM23YMM23XMM23
ZMM24YMM24XMM24
ZMM25YMM25XMM25
ZMM26YMM26XMM26
ZMM27YMM27XMM27
ZMM28YMM28XMM28
ZMM29YMM29XMM29
ZMM30YMM30XMM30
ZMM31YMM31XMM31

AVX introduces a three-operand SIMD instruction format called VEX coding scheme, where the destination register is distinct from the two source operands. For example, an SSE instruction using the conventional two-operand form aa + b can now use a non-destructive three-operand form ca + b, preserving both source operands. Originally, AVX's three-operand format was limited to the instructions with SIMD operands (YMM), and did not include instructions with general purpose registers (e.g. EAX). It was later used for coding new instructions on general purpose registers in later extensions, such as BMI. VEX coding is also used for instructions operating on the k0-k7 mask registers that were introduced with AVX-512.

The alignment requirement of SIMD memory operands is relaxed. [5] Unlike their non-VEX coded counterparts, most VEX coded vector instructions no longer require their memory operands to be aligned to the vector size. Notably, the VMOVDQA instruction still requires its memory operand to be aligned.

The new VEX coding scheme introduces a new set of code prefixes that extends the opcode space, allows instructions to have more than two operands, and allows SIMD vector registers to be longer than 128 bits. The VEX prefix can also be used on the legacy SSE instructions giving them a three-operand form, and making them interact more efficiently with AVX instructions without the need for VZEROUPPER and VZEROALL.

The AVX instructions support both 128-bit and 256-bit SIMD. The 128-bit versions can be useful to improve old code without needing to widen the vectorization, and avoid the penalty of going from SSE to AVX, they are also faster on some early AMD implementations of AVX. This mode is sometimes known as AVX-128. [6]

New instructions

These AVX instructions are in addition to the ones that are 256-bit extensions of the legacy 128-bit SSE instructions; most are usable on both 128-bit and 256-bit operands.

InstructionDescription
VBROADCASTSS, VBROADCASTSD, VBROADCASTF128Copy a 32-bit, 64-bit or 128-bit memory operand to all elements of a XMM or YMM vector register.
VINSERTF128Replaces either the lower half or the upper half of a 256-bit YMM register with the value of a 128-bit source operand. The other half of the destination is unchanged.
VEXTRACTF128Extracts either the lower half or the upper half of a 256-bit YMM register and copies the value to a 128-bit destination operand.
VMASKMOVPS, VMASKMOVPDConditionally reads any number of elements from a SIMD vector memory operand into a destination register, leaving the remaining vector elements unread and setting the corresponding elements in the destination register to zero. Alternatively, conditionally writes any number of elements from a SIMD vector register operand to a vector memory operand, leaving the remaining elements of the memory operand unchanged. On the AMD Jaguar processor architecture, this instruction with a memory source operand takes more than 300 clock cycles when the mask is zero, in which case the instruction should do nothing. This appears to be a design flaw. [7]
VPERMILPS, VPERMILPDPermute In-Lane. Shuffle the 32-bit or 64-bit vector elements of one input operand. These are in-lane 256-bit instructions, meaning that they operate on all 256 bits with two separate 128-bit shuffles, so they can not shuffle across the 128-bit lanes. [8]
VPERM2F128Shuffle the four 128-bit vector elements of two 256-bit source operands into a 256-bit destination operand, with an immediate constant as selector.
VTESTPS, VTESTPDPacked bit test of the packed single-precision or double-precision floating-point sign bits, setting or clearing the ZF flag based on AND and CF flag based on ANDN.
VZEROALLSet all YMM registers to zero and tag them as unused. Used when switching between 128-bit use and 256-bit use.
VZEROUPPERSet the upper half of all YMM registers to zero. Used when switching between 128-bit use and 256-bit use.

CPUs with AVX

Issues regarding compatibility between future Intel and AMD processors are discussed under XOP instruction set.

Compiler and assembler support

Operating system support

AVX adds new register-state through the 256-bit wide YMM register file, so explicit operating system support is required to properly save and restore AVX's expanded registers between context switches. The following operating system versions support AVX:

Advanced Vector Extensions 2

Advanced Vector Extensions 2 (AVX2), also known as Haswell New Instructions, [24] is an expansion of the AVX instruction set introduced in Intel's Haswell microarchitecture. AVX2 makes the following additions:

Sometimes three-operand fused multiply-accumulate (FMA3) extension is considered part of AVX2, as it was introduced by Intel in the same processor microarchitecture. This is a separate extension using its own CPUID flag and is described on its own page and not below.

New instructions

InstructionDescription
VBROADCASTSS, VBROADCASTSDCopy a 32-bit or 64-bit register operand to all elements of a XMM or YMM vector register. These are register versions of the same instructions in AVX1. There is no 128-bit version, but the same effect can be simply achieved using VINSERTF128.
VPBROADCASTB, VPBROADCASTW, VPBROADCASTD, VPBROADCASTQCopy an 8, 16, 32 or 64-bit integer register or memory operand to all elements of a XMM or YMM vector register.
VBROADCASTI128Copy a 128-bit memory operand to all elements of a YMM vector register.
VINSERTI128Replaces either the lower half or the upper half of a 256-bit YMM register with the value of a 128-bit source operand. The other half of the destination is unchanged.
VEXTRACTI128Extracts either the lower half or the upper half of a 256-bit YMM register and copies the value to a 128-bit destination operand.
VGATHERDPD, VGATHERQPD, VGATHERDPS, VGATHERQPS Gathers single or double precision floating point values using either 32 or 64-bit indices and scale.
VPGATHERDD, VPGATHERDQ, VPGATHERQD, VPGATHERQQGathers 32 or 64-bit integer values using either 32 or 64-bit indices and scale.
VPMASKMOVD, VPMASKMOVQConditionally reads any number of elements from a SIMD vector memory operand into a destination register, leaving the remaining vector elements unread and setting the corresponding elements in the destination register to zero. Alternatively, conditionally writes any number of elements from a SIMD vector register operand to a vector memory operand, leaving the remaining elements of the memory operand unchanged.
VPERMPS, VPERMDShuffle the eight 32-bit vector elements of one 256-bit source operand into a 256-bit destination operand, with a register or memory operand as selector.
VPERMPD, VPERMQShuffle the four 64-bit vector elements of one 256-bit source operand into a 256-bit destination operand, with a register or memory operand as selector.
VPERM2I128Shuffle (two of) the four 128-bit vector elements of two 256-bit source operands into a 256-bit destination operand, with an immediate constant as selector.
VPBLENDDDoubleword immediate version of the PBLEND instructions from SSE4.
VPSLLVD, VPSLLVQShift left logical. Allows variable shifts where each element is shifted according to the packed input.
VPSRLVD, VPSRLVQShift right logical. Allows variable shifts where each element is shifted according to the packed input.
VPSRAVDShift right arithmetically. Allows variable shifts where each element is shifted according to the packed input.

CPUs with AVX2

AVX-512

AVX-512 are 512-bit extensions to the 256-bit Advanced Vector Extensions SIMD instructions for x86 instruction set architecture proposed by Intel in July 2013. [3]

AVX-512 instructions are encoded with the new EVEX prefix. It allows 4 operands, 8 new 64-bit opmask registers, scalar memory mode with automatic broadcast, explicit rounding control, and compressed displacement memory addressing mode. The width of the register file is increased to 512 bits and total register count increased to 32 (registers ZMM0-ZMM31) in x86-64 mode.

AVX-512 consists of multiple instruction subsets, not all of which are meant to be supported by all processors implementing them. The instruction set consists of the following:

Only the core extension AVX-512F (AVX-512 Foundation) is required by all implementations, though all current implementations also support CD (conflict detection). All central processors with AVX-512 also support VL, DQ and BW. The ER, PF, 4VNNIW and 4FMAPS instruction set extensions are currently only implemented in Intel computing coprocessors.

The updated SSE/AVX instructions in AVX-512F use the same mnemonics as AVX versions; they can operate on 512-bit ZMM registers, and will also support 128/256 bit XMM/YMM registers (with AVX-512VL) and byte, word, doubleword and quadword integer operands (with AVX-512BW/DQ and VBMI). [26] :23

AVX-512 CPU compatibility table

Subset
F
CD
ER
PF
4FMAPS
4VNNIW
VPOPCNTDQ
VL
DQ
BW
IFMA
VBMI
VBMI2
BITALG
VNNI
BF16
VPCLMULQDQ
GFNI
VAES
VP2INTERSECT
FP16
Intel Knights Landing (2016)YesYesNo
Intel Knights Mill (2017)YesNo
Intel Skylake-SP, Skylake-X (2017)NoNoYesNo
Intel Cannon Lake (2018)YesNo
Intel Cascade Lake-SP (2019)NoYesNo
Intel Cooper Lake (2020)NoYesNo
Intel Ice Lake (2019)YesNoYesNo
Intel Tiger Lake (2020)YesNo
Intel Rocket Lake (2021)No
Intel Alder Lake (2021)Partial Note 1 Partial Note 1
AMD Zen 4 (2022)YesYesNo
Intel Sapphire Rapids (2023)NoYes
AMD Zen 5 (2024)YesNo

[28]

^Note 1 : Intel does not officially support AVX-512 family of instructions on the Alder Lake microprocessors. In early 2022, Intel began disabling in silicon (fusing off) AVX-512 in Alder Lake microprocessors to prevent customers from enabling AVX-512. [29] In older Alder Lake family CPUs with some legacy combinations of BIOS and microcode revisions, it was possible to execute AVX-512 family instructions when disabling all the efficiency cores which do not contain the silicon for AVX-512. [30] [31] [32]

Compilers supporting AVX-512

Assemblers supporting AVX-512

AVX-VNNI, AVX-IFMA

AVX-VNNI is a VEX-coded variant of the AVX512-VNNI instruction set extension. Similarly, AVX-IFMA is a VEX-coded variant of AVX512-IFMA. These extensions provide the same sets of operations as their AVX-512 counterparts, but are limited to 256-bit vectors and do not support any additional features of EVEX encoding, such as broadcasting, opmask registers or accessing more than 16 vector registers. These extensions allow support of VNNI and IFMA operations even when full AVX-512 support is not implemented in the processor.

CPUs with AVX-VNNI

CPUs with AVX-IFMA

AVX10

AVX10, announced in July 2023, [38] is a new, "converged" AVX instruction set. It addresses several issues of AVX-512, in particular that it is split into too many parts [39] (20 feature flags) and that it makes 512-bit vectors mandatory to support. AVX10 presents a simplified CPUID interface to test for instruction support, consisting of the AVX10 version number (indicating the set of instructions supported, with later versions always being a superset of an earlier one) and the available maximum vector length (256 or 512 bits). [40] A combined notation is used to indicate the version and vector length: for example, AVX10.2/256 indicates that a CPU is capable of the second version of AVX10 with a maximum vector width of 256 bits. [41]

The first and "early" version of AVX10, notated AVX10.1, will not introduce any instructions or encoding features beyond what is already in AVX-512 (specifically, in Intel Sapphire Rapids: AVX-512F, CD, VL, DQ, BW, IFMA, VBMI, VBMI2, BITALG, VNNI, GFNI, VPOPCNTDQ, VPCLMULQDQ, VAES, BF16, FP16). The second and "fully-featured" version, AVX10.2, introduces new features such as YMM embedded rounding and Suppress All Exception. For CPUs supporting AVX10 and 512-bit vectors, all legacy AVX-512 feature flags will remain set to facilitate applications supporting AVX-512 to continue using AVX-512 instructions. [41]

AVX10.1/512 was first released in Intel Granite Rapids [41] (Q3 2024) and AVX10.2/512 will be available in Diamond Rapids. [42]

APX

APX is a new extension. It is not focused on vector computation, but provides RISC-like extensions to the x86-64 architecture by doubling the number of general-purpose registers to 32 and introducing three-operand instruction formats. AVX is only tangentially affected as APX introduces extended operands. [43] [44]

Applications

Software

Downclocking

Since AVX instructions are wider, they consume more power and generate more heat. Executing heavy AVX instructions at high CPU clock frequencies may affect CPU stability due to excessive voltage droop during load transients. Some Intel processors have provisions to reduce the Turbo Boost frequency limit when such instructions are being executed. This reduction happens even if the CPU hasn't reached its thermal and power consumption limits. On Skylake and its derivatives, the throttling is divided into three levels: [66] [67]

The frequency transition can be soft or hard. Hard transition means the frequency is reduced as soon as such an instruction is spotted; soft transition means that the frequency is reduced only after reaching a threshold number of matching instructions. The limit is per-thread. [66]

In Ice Lake, only two levels persist: [68]

Rocket Lake processors do not trigger frequency reduction upon executing any kind of vector instructions regardless of the vector size. [68] However, downclocking can still happen due to other reasons, such as reaching thermal and power limits.

Downclocking means that using AVX in a mixed workload with an Intel processor can incur a frequency penalty. Avoiding the use of wide and heavy instructions help minimize the impact in these cases. AVX-512VL allows for using 256-bit or 128-bit operands in AVX-512 instructions, making it a sensible default for mixed loads. [69]

On supported and unlocked variants of processors that down-clock, the clock ratio reduction offsets (typically called AVX and AVX-512 offsets) are adjustable and may be turned off entirely (set to 0x) via Intel's Overclocking / Tuning utility or in BIOS if supported there. [70]

See also

Related Research Articles

x86 Family of instruction set architectures

x86 is a family of complex instruction set computer (CISC) instruction set architectures initially developed by Intel based on the 8086 microprocessor and its 8-bit-external-bus variant, the 8088. The 8086 was introduced in 1978 as a fully 16-bit extension of 8-bit Intel's 8080 microprocessor, with memory segmentation as a solution for addressing more memory than can be covered by a plain 16-bit address. The term "x86" came into being because the names of several successors to Intel's 8086 processor end in "86", including the 80186, 80286, 80386 and 80486. Colloquially, their names were "186", "286", "386" and "486".

<span class="mw-page-title-main">Single instruction, multiple data</span> Type of parallel processing

Single instruction, multiple data (SIMD) is a type of parallel processing in Flynn's taxonomy. SIMD can be internal and it can be directly accessible through an instruction set architecture (ISA), but it should not be confused with an ISA. SIMD describes computers with multiple processing elements that perform the same operation on multiple data points simultaneously.

<span class="mw-page-title-main">MMX (instruction set)</span> Instruction set designed by Intel

MMX is a single instruction, multiple data (SIMD) instruction set architecture designed by Intel, introduced on January 8, 1997 with its Pentium P5 (microarchitecture) based line of microprocessors, named "Pentium with MMX Technology". It developed out of a similar unit introduced on the Intel i860, and earlier the Intel i750 video pixel processor. MMX is a processor supplementary capability that is supported on IA-32 processors by Intel and other vendors as of 1997. AMD also added MMX instruction set in its K6 processor.

In computing, Streaming SIMD Extensions (SSE) is a single instruction, multiple data (SIMD) instruction set extension to the x86 architecture, designed by Intel and introduced in 1999 in its Pentium III series of central processing units (CPUs) shortly after the appearance of Advanced Micro Devices (AMD's) 3DNow!. SSE contains 70 new instructions, most of which work on single precision floating-point data. SIMD instructions can greatly increase performance when exactly the same operations are to be performed on multiple data objects. Typical applications are digital signal processing and graphics processing.

x86-64 64-bit version of x86 architecture

x86-64 is a 64-bit version of the x86 instruction set, first announced in 1999. It introduced two new modes of operation, 64-bit mode and compatibility mode, along with a new 4-level paging mode.

SSE2 is one of the Intel SIMD processor supplementary instruction sets introduced by Intel with the initial version of the Pentium 4 in 2000. SSE2 instructions allow the use of XMM (SIMD) registers on x86 instruction set architecture processors. These registers can load up to 128 bits of data and perform instructions, such as vector addition and multiplication, simultaneously.

The x86 instruction set refers to the set of instructions that x86-compatible microprocessors support. The instructions are usually part of an executable program, often stored as a computer file and executed on the processor.

SSE4 is a SIMD CPU instruction set used in the Intel Core microarchitecture and AMD K10 (K8L). It was announced on September 27, 2006, at the Fall 2006 Intel Developer Forum, with vague details in a white paper; more precise details of 47 instructions became available at the Spring 2007 Intel Developer Forum in Beijing, in the presentation. SSE4 extended the SSE3 instruction set which was released in early 2004. All software using previous Intel SIMD instructions are compatible with modern microprocessors supporting SSE4 instructions. All existing software continues to run correctly without modification on microprocessors that incorporate SSE4, as well as in the presence of existing and new applications that incorporate SSE4.

The SSE5 was a SIMD instruction set extension proposed by AMD on August 30, 2007 as a supplement to the 128-bit SSE core instructions in the AMD64 architecture.

The XOP instruction set, announced by AMD on May 1, 2009, is an extension to the 128-bit SSE core instructions in the x86 and AMD64 instruction set for the Bulldozer processor core, which was released on October 12, 2011. However AMD removed support for XOP from Zen (microarchitecture) onward.

The VEX prefix and VEX coding scheme are an extension to the IA-32 and x86-64 instruction set architecture for microprocessors from Intel, AMD and others.

The FMA instruction set is an extension to the 128 and 256-bit Streaming SIMD Extensions instructions in the x86 microprocessor instruction set to perform fused multiply–add (FMA) operations. There are two variants:

Carry-less Multiplication (CLMUL) is an extension to the x86 instruction set used by microprocessors from Intel and AMD which was proposed by Intel in March 2008 and made available in the Intel Westmere processors announced in early 2010. Mathematically, the instruction implements multiplication of polynomials over the finite field GF(2) where the bitstring represents the polynomial . The CLMUL instruction also allows a more efficient implementation of the closely related multiplication of larger finite fields GF(2k) than the traditional instruction set.

AVX-512 are 512-bit extensions to the 256-bit Advanced Vector Extensions SIMD instructions for x86 instruction set architecture (ISA) proposed by Intel in July 2013, and first implemented in the 2016 Intel Xeon Phi x200, and then later in a number of AMD and other Intel CPUs. AVX-512 consists of multiple extensions that may be implemented independently. This policy is a departure from the historical requirement of implementing the entire instruction block. Only the core extension AVX-512F is required by all AVX-512 implementations.

Intel SHA Extensions are a set of extensions to the x86 instruction set architecture which support hardware acceleration of Secure Hash Algorithm (SHA) family. It was specified in 2013. Instructions for SHA-512 was introduced in Arrow Lake and Lunar Lake in 2024.

The EVEX prefix and corresponding coding scheme is an extension to the 32-bit x86 (IA-32) and 64-bit x86-64 (AMD64) instruction set architecture. EVEX is based on, but should not be confused with the MVEX prefix used by the Knights Corner processor.

Bit manipulation instructions sets are extensions to the x86 instruction set architecture for microprocessors from Intel and AMD. The purpose of these instruction sets is to improve the speed of bit manipulation. All the instructions in these sets are non-SIMD and operate only on general-purpose registers.

The x86 instruction set has several times been extended with SIMD instruction set extensions. These extensions, starting from the MMX instruction set extension introduced with Pentium MMX in 1997, typically define sets of wide registers and instructions that subdivide these registers into fixed-size lanes and perform a computation for each lane in parallel.

References

  1. Kanter, David (September 25, 2010). "Intel's Sandy Bridge Microarchitecture". www.realworldtech.com. Retrieved February 17, 2018.
  2. Hruska, Joel (October 24, 2011). "Analyzing Bulldozer: Why AMD's chip is so disappointing - Page 4 of 5 - ExtremeTech". ExtremeTech. Retrieved February 17, 2018.
  3. 1 2 3 4 5 James Reinders (July 23, 2013), AVX-512 Instructions, Intel , retrieved August 20, 2013
  4. "Intel Xeon Phi Processor 7210 (16GB, 1.30 GHz, 64 core) Product Specifications". Intel ARK (Product Specs). Retrieved March 16, 2018.
  5. "14.9". Intel 64 and IA-32 Architectures Software Developer's Manual Volume 1: Basic Architecture (PDF) (-051US ed.). Intel Corporation. p. 349. Retrieved August 23, 2014. Memory arguments for most instructions with VEX prefix operate normally without causing #GP(0) on any byte-granularity alignment (unlike Legacy SSE instructions).
  6. "i386 and x86-64 Options - Using the GNU Compiler Collection (GCC)" . Retrieved February 9, 2014.
  7. "The microarchitecture of Intel, AMD and VIA CPUs: An optimization guide for assembly programmers and compiler makers" (PDF). Retrieved October 17, 2016.
  8. "Chess programming AVX2". Archived from the original on July 10, 2017. Retrieved October 17, 2016.
  9. "Intel Offers Peek at Nehalem and Larrabee". ExtremeTech. March 17, 2008.
  10. 1 2 "Intel® Celeron® 6305 Processor (4M Cache, 1.80 GHz, with IPU) Product Specifications". ark.intel.com. Retrieved November 10, 2020.
  11. Butler, Michael; Barnes, Leslie; Das Sarma, Debjit; Gelinas, Bob (March–April 2011). "Bulldozer: An Approach to Multithreaded Compute Performance" (PDF). IEEE Micro . 31 (2): 6–15. doi:10.1109/MM.2011.23. S2CID   28236214. Archived from the original (PDF) on May 19, 2024.
  12. "What's New - RAD Studio". docwiki.embarcadero.com. Retrieved September 17, 2021.
  13. "GAS Changes". sourceware.org. Retrieved May 3, 2024.
  14. 1 2 "NASM - The Netwide Assembler, Appendix C: NASM Version History". nasm.us. Retrieved May 3, 2024.
  15. "YASM 0.7.0 Release Notes". yasm.tortall.net.
  16. Add support for the extended FPU states on amd64, both for native 64bit and 32bit ABIs, svnweb.freebsd.org, January 21, 2012, retrieved January 22, 2012
  17. "FreeBSD 9.1-RELEASE Announcement" . Retrieved May 20, 2013.
  18. x86: add linux kernel support for YMM state , retrieved July 13, 2009
  19. Linux 2.6.30 - Linux Kernel Newbies , retrieved July 13, 2009
  20. Twitter , retrieved June 23, 2010
  21. "Devs are making progress getting macOS Ventura to run on unsupported, decade-old Macs". August 23, 2022.
  22. Add support for saving/restoring FPU state using the XSAVE/XRSTOR. , retrieved March 25, 2015
  23. Floating-Point Support for 64-Bit Drivers , retrieved December 6, 2009
  24. Haswell New Instruction Descriptions Now Available, Software.intel.com, retrieved January 17, 2012
  25. 1 2 3 James Reinders (July 17, 2014). "Additional AVX-512 instructions". Intel . Retrieved August 3, 2014.
  26. 1 2 "Intel Architecture Instruction Set Extensions Programming Reference" (PDF). Intel . Retrieved January 29, 2014.
  27. 1 2 3 4 5 6 7 "Intel® Architecture Instruction Set Extensions and Future Features Programming Reference". Intel. Retrieved October 16, 2017.
  28. "Intel® Software Development Emulator | Intel® Software". software.intel.com. Retrieved June 11, 2016.
  29. Alcorn, Paul (March 2, 2022). "Intel Nukes Alder Lake's AVX-512 Support, Now Fuses It Off in Silicon". Tom's Hardware. Retrieved March 7, 2022.
  30. Cutress, Ian; Frumusanu, Andrei (August 19, 2021). "Intel Architecture Day 2021: Alder Lake, Golden Cove, and Gracemont Detailed". AnandTech. Retrieved August 25, 2021.
  31. Alcorn, Paul (August 19, 2021). "Intel Architecture Day 2021: Alder Lake Chips, Golden Cove and Gracemont Cores". Tom's Hardware. Retrieved August 21, 2021.
  32. Cutress, Ian; Frumusanu, Andrei. "The Intel 12th Gen Core i9-12900K Review: Hybrid Performance Brings Hybrid Complexity". www.anandtech.com. Retrieved November 5, 2021.
  33. "LLVM 3.9 Release Notes — LLVM 3.9 documentation". releases.llvm.org. Retrieved April 3, 2017.
  34. "GCC 4.9 Release Series — Changes, New Features, and Fixes  GNU Project - Free Software Foundation (FSF)". gcc.gnu.org. Retrieved April 3, 2017.
  35. "Intel® Parallel Studio XE 2015 Composer Edition C++ Release Notes | Intel® Software". software.intel.com. Retrieved April 3, 2017.
  36. "Microsoft Visual Studio 2017 Supports Intel® AVX-512". July 11, 2017.
  37. "AMD Zen 5 Compiler Support Posted For GCC - Confirms New AVX Features & More". www.phoronix.com. Retrieved February 10, 2024.
  38. Bonshor, Gavin (July 25, 2023). "Intel Unveils AVX10 and APX Instruction Sets: Unifying AVX-512 For Hybrid Architectures". AnandTech . Retrieved August 21, 2024.
  39. Mann, Tobias (August 15, 2023). "Intel's AVX10 promises benefits of AVX-512 without baggage". www.theregister.com. Retrieved August 20, 2023.
  40. "The Converged Vector ISA: Intel® Advanced Vector Extensions 10 Technical Paper". Intel.
  41. 1 2 3 "Intel® Advanced Vector Extensions 10 (Intel® AVX10) Architecture Specification". Intel.
  42. Larabel, Michael (October 23, 2024). "Intel Preps GCC Compiler For New AMX & ISA Features Ahead Of Diamond Rapids". Phoronix. Retrieved October 23, 2024.
  43. "Intel® Advanced Performance Extensions (Intel® APX) Architecture Specification". Intel.
  44. Robinson, Dan (July 26, 2023). "Intel discloses x86 and vector instructions for future chips". www.theregister.com. Retrieved August 20, 2023.
  45. "Linux RAID". LWN. February 17, 2013. Archived from the original on April 15, 2013.
  46. "Comparison of BSAFE cryptographic library implementations". July 25, 2023.
  47. "Improving OpenSSL Performance". May 26, 2015. Retrieved February 28, 2017.
  48. "OpenSSL 3.0.0 release notes". GitHub . September 7, 2021.
  49. Jaroš, Milan; Strakoš, Petr; Říha, Lubomír (May 28, 2022). "Rendering in Blender using AVX-512 Vectorization" (PDF). Intel eXtreme Performance Users Group. Technical University of Ostrava . Retrieved October 28, 2022.
  50. "MASSIVE X Requires AVX Compatible Processor". Native Instruments. Retrieved November 29, 2019.
  51. "dav1d: performance and completion of the first release". November 21, 2018. Retrieved November 22, 2018.
  52. "dav1d 0.6.0 release notes". March 6, 2020.
  53. "SVT-AV1 0.7.0 release notes". September 26, 2019.
  54. "ArcGIS Data Store 11.2 System Requirements". ArcGIS Enterprise. Retrieved January 24, 2024.
  55. "Prime95 release notes" . Retrieved July 10, 2022.
  56. "Einstein@Home Applications".
  57. "FAQ, Helios". Helios. Retrieved July 5, 2021.
  58. "Tensorflow 1.6". GitHub .
  59. New in Version 19.0 – EmEditor (Text Editor)
  60. "Hardware requirements for Microsoft Teams". Microsoft. Retrieved April 17, 2020.
  61. "Reduce background noise in Teams meetings". Microsoft Support. Retrieved January 5, 2021.
  62. Langdale, Geoff; Lemire, Daniel (2019). "Parsing Gigabytes of JSON per Second". The VLDB Journal. 28 (6): 941–960. arXiv: 1902.08318 . doi:10.1007/s00778-019-00578-5. S2CID   67856679.
  63. "simdjson 2.1.0 release notes". GitHub . June 30, 2022.
  64. Larabel, Michael (October 6, 2023). "OpenJDK Merges Intel's x86-simd-sort For Speeding Up Data Sorting 7~15x". Phoronix.
  65. Larabel, Michael (July 7, 2022). "Tesseract OCR 5.2 Engine Finds Success With AVX-512F". Phoronix.
  66. 1 2 Lemire, Daniel (September 7, 2018). "AVX-512: when and how to use these new instructions". Daniel Lemire's blog.
  67. BeeOnRope. "SIMD instructions lowering CPU frequency". Stack Overflow.
  68. 1 2 Downs, Travis (August 19, 2020). "Ice Lake AVX-512 Downclocking". Performance Matters blog.
  69. "x86 - AVX 512 vs AVX2 performance for simple array processing loops". Stack Overflow.
  70. "Intel® Extreme Tuning Utility (Intel® XTU) Guide to Overclocking : Advanced Tuning". Intel. Retrieved July 18, 2021. See image in linked section, where AVX2 ratio has been set to 0.