Graphcore

Last updated
Graphcore Limited
Type Private
Industry Semiconductors
Founded2016;7 years ago (2016)
Founders
  • Nigel Toon
  • Simon Knowles
Headquarters,
Key people
  • Nigel Toon (CEO)
  • Simon Knowles (CTO)
ProductsIPU, Poplar
RevenueUS$2.7 million (2022) [1]
US$−205 million (2022) [1]
Number of employees
494 (2023) [1]
Website www.graphcore.ai

Graphcore Limited is a British semiconductor company that develops accelerators for AI and machine learning. It has introduced a massively parallel Intelligence Processing Unit (IPU) that holds the complete machine learning model inside the processor. [2]

Contents

History

Graphcore was founded in 2016 by Simon Knowles and Nigel Toon. [3]

In the autumn of 2016, Graphcore secured a first funding round led by Robert Bosch Venture Capital. Other backers include Samsung, Amadeus Capital Partners, C4 Ventures, Draper Esprit, Foundation Capital, and Pitango. [4] [5]

In July 2017, Graphcore secured a round B funding led by Atomico, [6] which was followed a few months later by $50 million in funding from Sequoia Capital. [7]

In December 2018, Graphcore closed its series D with $200 million raised at a $1.7 billion valuation, making the company a unicorn. Investors included Microsoft, Samsung and Dell Technologies. [8]

On 13 November 2019, Graphcore announced that their Graphcore C2 IPUs are available for preview on Microsoft Azure. [9]

Meta Platforms acquired the AI networking technology team from Graphcore in early 2023. [10]

Products

In 2016, Graphcore announced the world's first graph tool chain designed for machine intelligence called Poplar Software Stack. [11] [12] [13]

In July 2017, Graphcore announced their first chip, called the Colossus GC2, a "16 nm massively parallel, mixed-precision floating point processor", first available in 2018. [14] [15] Packaged with two chips on a single PCI Express card called the Graphcore C2 IPU (an Intelligence Processing Unit), it is stated to perform the same role as a GPU in conjunction with standard machine learning frameworks such as TensorFlow. [14] The device relies on scratchpad memory for its performance rather than traditional cache hierarchies. [16]

In July 2020, Graphcore presented their second generation processor called GC200 built in TSMC's 7nm FinFET manufacturing process. GC200 is a 59 billion transistor, 823 square-millimeter integrated circuit with 1,472 computational cores and 900 Mbyte of local memories. [17] In 2022, Graphcore and TSMC presented the Bow IPU, a 3D package of a GC200 die bonded face to face to a power-delivery die that allows for higher clock rate at lower core voltage. [18] Graphcore aims at a Good machine, named after I.J. Good, enabling AI models with more parameters than the human brain has synapses. [18]

Release dateProductProcess nodeCoresThreadsTransistorsteraFLOPS (FP16)
July 2017Colossus™ MK1 - GC2 IPU16 nm TSMC12167296 ?~100-125 [19]
July 2020Colossus™ MK2 - GC200 IPU7 nm TSMC1472883259 billion~250-280 [20]
Colossus™ MK3~500 [21]

Both the older and newer chips can use 6 threads per tile (for a total of 7,296 and 8,832 threads, respectively) "MIMD (Multiple Instruction, Multiple Data) parallelism and has distributed, local memory as its only form of memory on the device" (except for registers). The older GC2 chip has 256 KiB per tile while the newer GC200 chip has about 630 KiB per tile that are arranged into islands (4 tiles per island), [22] that are arranged into columns, and latency is best within tile. The IPU uses IEEE FP16, with stochastic rounding, and also single-precision FP32, at lower performance. [23] Code and data executed locally must fit in a tile, but with message-passing, all on-chip or off-chip memory can be used, and software for AI makes it transparently possible, e.g. has PyTorch support.

See also

Related Research Articles

IPU may refer to:

<span class="mw-page-title-main">Xilinx</span> American technology company

Xilinx, Inc. was an American technology and semiconductor company that primarily supplied programmable logic devices. The company is known for inventing the first commercially viable field-programmable gate array (FPGA) and creating the first fabless manufacturing model.

The transistor count is the number of transistors in an electronic device. It is the most common measure of integrated circuit complexity. The rate at which MOS transistor counts have increased generally follows Moore's law, which observes that transistor count doubles approximately every two years. However, being directly proportional to the area of a chip, transistor count does not represent how advanced the corresponding manufacturing technology is: a better indication of this is transistor density.

The 22 nm node is the process step following 32 nm in CMOS MOSFET semiconductor device fabrication. The typical half-pitch for a memory cell using the process is around 22 nm. It was first demonstrated by semiconductor companies for use in RAM memory in 2008. In 2010, Toshiba began shipping 24 nm flash memory chips, and Samsung Electronics began mass-producing 20 nm flash memory chips. The first consumer-level CPU deliveries using a 22 nm process started in April 2012 with the Intel Ivy Bridge processors.

<span class="mw-page-title-main">CUDA</span> Parallel computing platform and programming model

CUDA is a proprietary and closed source parallel computing platform and application programming interface (API) that allows software to use certain types of graphics processing units (GPUs) for general purpose processing, an approach called general-purpose computing on GPUs (GPGPU). CUDA is a software layer that gives direct access to the GPU's virtual instruction set and parallel computational elements, for the execution of compute kernels.

The 14 nm process refers to the MOSFET technology node that is the successor to the 22 nm node. The 14 nm was so named by the International Technology Roadmap for Semiconductors (ITRS). Until about 2011, the node following 22 nm was expected to be 16 nm. All 14 nm nodes use FinFET technology, a type of multi-gate MOSFET technology that is a non-planar evolution of planar silicon CMOS technology.

Achronix Semiconductor Corporation is an American fabless semiconductor company based in Santa Clara, California with an additional R&D facility in Bangalore, India, and an additional sales office in Shenzhen, China. Achronix is a diversified fabless semiconductor company that sells FPGA products, embedded FPGA (eFPGA) products, system-level products and supporting design tools. Achronix was founded in 2004 in Ithaca, New York based on technology licensed from Cornell University. In 2006, Achronix moved its headquarters to Silicon Valley.

<span class="mw-page-title-main">Volta (microarchitecture)</span> GPU microarchitecture by Nvidia

Volta is the codename, but not the trademark, for a GPU microarchitecture developed by Nvidia, succeeding Pascal. It was first announced on a roadmap in March 2013, although the first product was not announced until May 2017. The architecture is named after 18th–19th century Italian chemist and physicist Alessandro Volta. It was NVIDIA's first chip to feature Tensor Cores, specially designed cores that have superior deep learning performance over regular CUDA cores. The architecture is produced with TSMC's 12 nm FinFET process. The Ampere microarchitecture is the successor to Volta.

Annapurna Labs is an Israeli microelectronics company. Since January 2015 it has been a wholly-owned subsidiary of Amazon.com. Amazon reportedly acquired the company for its Amazon Web Services division for US$350–370M.

A vision processing unit (VPU) is an emerging class of microprocessor; it is a specific type of AI accelerator, designed to accelerate machine vision tasks.

<span class="mw-page-title-main">Tensor Processing Unit</span> AI accelerator ASIC by Google

Tensor Processing Unit (TPU) is an AI accelerator application-specific integrated circuit (ASIC) developed by Google for neural network machine learning, using Google's own TensorFlow software. Google began using TPUs internally in 2015, and in 2018 made them available for third party use, both as part of its cloud infrastructure and by offering a smaller version of the chip for sale.

<span class="mw-page-title-main">AMD Instinct</span> Brand name by AMD; professional GPUs for high-performance-computing, machine learning

AMD Instinct is AMD's brand of professional GPUs. It replaced AMD's FirePro S brand in 2016. Compared to the Radeon brand of mainstream consumer/gamer products, the Instinct product line is intended to accelerate deep learning, artificial neural network, and high-performance computing/GPGPU applications.

<span class="mw-page-title-main">SiFive</span> Fabless semiconductor company providing RISC-V processors

SiFive, Inc. is a fabless semiconductor company and provider of commercial RISC-V processor IP and silicon chips based on the RISC-V instruction set architecture (ISA). SiFive's products include cores, SoCs, IPs, and development boards.

<span class="mw-page-title-main">Apple A11</span> System on a chip (SoC) designed by Apple Inc.

The Apple A11 Bionic is a 64-bit ARM-based system on a chip (SoC), designed by Apple Inc. and manufactured by TSMC. It first appeared in the iPhone 8 and 8 Plus, and iPhone X which were introduced on September 12, 2017. Apple states that the two high-performance cores are 25% faster than the Apple A10's and the four high-efficiency cores are up to 70% faster than the two corresponding cores in the A10. The A11 Bionic chip was discontinued on April 15, 2020, following the discontinuation of the iPhone 8 and 8 Plus. The latest software update for the iPhone 8 & 8 Plus and iPhone X using this chip was iOS 16.7, released on September 21, 2023.

The Pixel Visual Core (PVC) is a series of ARM-based system in package (SiP) image processors designed by Google. The PVC is a fully programmable image, vision and AI multi-core domain-specific architecture (DSA) for mobile devices and in future for IoT. It first appeared in the Google Pixel 2 and 2 XL which were introduced on October 19, 2017. It has also appeared in the Google Pixel 3 and 3 XL. Starting with the Pixel 4, this chip was replaced with the Pixel Neural Core.

In semiconductor manufacturing, the 3 nm process is the next die shrink after the 5 nanometer MOSFET technology node. South Korean chipmaker Samsung started shipping its 3 nm gate all around (GAA) process, named 3GAA, in mid-2022. On 29 December 2022, Taiwanese chip manufacturer TSMC announced that volume production using its 3 nm semiconductor node termed N3 is under way with good yields. An enhanced 3 nm chip process called N3E may have started production in 2023. American manufacturer Intel planned to start 3 nm production in 2023.

<span class="mw-page-title-main">Ampere Computing</span> American fabless semiconductor company

Ampere Computing LLC is an American fabless semiconductor company based in Santa Clara, California that develops processes for servers operating in large scale environments. Ampere also has offices in: Portland, Oregon; Taipei, Taiwan; Raleigh, North Carolina; Bangalore, India; Warsaw, Poland; and Ho Chi Minh City, Vietnam.

<span class="mw-page-title-main">Cerebras</span> American semiconductor company

Cerebras Systems is an American artificial intelligence company with offices in Sunnyvale and San Diego, Toronto, Tokyo and Bangalore, India. Cerebras builds computer systems for complex artificial intelligence deep learning applications.

Meta AI is an artificial intelligence laboratory that belongs to Meta Platforms Inc. Meta AI intends to develop various forms of artificial intelligence, improving augmented and artificial reality technologies. Meta AI is an academic research laboratory focused on generating knowledge for the AI community. This is in contrast to Facebook's Applied Machine Learning (AML) team, which focuses on practical applications of its products.

Shanghai Biren Intelligent Technology Co. is a Chinese fabless semiconductor design company. The company was founded in 2019 by Lingjie Xu and others, all of whom were previously employed at NVIDIA or Alibaba. Biren has advertised two general-purpose graphics processing units (GPGPUs), the BR100 and BR104. Both cards are aimed at artificial intelligence and high-performance computing.

References

  1. 1 2 3 Cherney, Max A. (5 October 2023). "Losses widen, cash needed at chip startup Graphcore, an Nvidia rival, filing shows". Reuters.
  2. Peter Clarke (2016-11-01). "AI Chip Startup Shares Insights: "Very large" FinFET chip in the works at TSMC". eetimes. Retrieved 2017-08-02.
  3. Jolly, Jasper (2020-12-29). "UK chipmaker Graphcore valued at $2.8bn after it raises $222m". The Guardian .
  4. Arjun Kharpal (2016-10-31). "AI chipmaker Graphcore raises $30 million to take on Intel". CNBC. Retrieved 2017-07-31.
  5. Madhumita Murgia (2016-10-31). "UK chip start-up Graphcore raises £30m for take on AI giants". Financial Times. Retrieved 2017-08-02.
  6. Jeremy Kahn and Ian King (2017-07-20). "U.K. Chip Designer Graphcore Gets $30 Million to Fund Expansion". Bloomberg. Retrieved 2017-07-31.
  7. Lynley, Matthew (2017-11-12). "Graphcore raises $50M amid a flurry of AI chip activity". TechCrunch. Retrieved 2017-12-07.
  8. "AI chip startup Graphcore closes $200M Series D, adds BMW and Microsoft as strategic investors". TechCrunch. 18 December 2018. Retrieved 2018-12-19.
  9. Toon, Nigel. "Microsoft and Graphcore collaborate to accelerate Artificial Intelligence". www.graphcore.ai. Retrieved 2019-11-16.
  10. Paul, Katie (5 May 2023). "Meta Platforms scoops up AI networking chip team from Graphcore". Reuters.
  11. Fyles, Matt. "Inside an AI 'brain' - What does machine learning look like?". www.graphcore.ai. Retrieved 2019-11-16.
  12. Doherty, Sally. "Introducing Poplar® - our IPU-Processor software at NeurIPS". www.graphcore.ai. Retrieved 2019-11-16.
  13. Fyles, Matt. "Graph computing for machine intelligence with Poplar™". www.graphcore.ai. Retrieved 2019-11-16.
  14. 1 2 Trader, Tiffany (2017-07-20). "Graphcore Readies Launch of 16nm Colossus-IPU Chip". hpcwire.com. HPC Wire. Retrieved 2017-12-11.
  15. Lucchesi, Ray (2018-11-19). "New GraphCore GC2 chips with 2PFlop performance in a Dell Server". silvertonconsulting.com. Silverton Consulting. Retrieved 2018-12-16.
  16. Citadel High Performance Computing R&D Team (2019). "Dissecting the Graphcore IPU Architecture via Microbenchmarking" (PDF).
  17. "Graphcore Introducing 2nd Generation IPU Systems For AI At Scale" . Retrieved 2020-08-09.
  18. 1 2 Timothy Prickett Morgan: GraphCore Goes Full 3D With AI Chips. The Next Platform, March 3, 2022.
  19. Kennedy, Patrick (2019-06-07). "Hands-on With a Graphcore C2 IPU PCIe Card at Dell Tech World". ServeTheHome. Retrieved 2023-06-26.
  20. Ltd, Graphcore. "IPU Processors". www.graphcore.ai. Retrieved 2023-06-26.
  21. "ScalAH22: 13th Workshop on Latest Advances in Scalable Algorithms for Large-Scale Heterogeneous Systems". www.csm.ornl.gov. Retrieved 2023-06-26.
  22. Jia, Zhe; Tillman, Blake; Maggioni, Marco; Daniele Paolo Scarpazza (2019). "Dissecting theGraphcore IPUArchitecturevia Microbenchmarking". arXiv: 1912.03413 [cs.DC].
  23. "THE GRAPHCORE SECOND GENERATION IPU" (PDF).

51°27′19.0″N2°35′33.3″W / 51.455278°N 2.592583°W / 51.455278; -2.592583