Arctic Region Supercomputing Center

Last updated
Arctic Region Supercomputing Center
ARSC Logo.png
Established1993
Affiliation University of Alaska Fairbanks
DirectorDr. Gregory Newby
Administrative staff
30
Location,
USA
Website arsc.edu
C.T. Elvey Building, former location of ARSC Elvey Building former home of ARSC (46793543565).jpg
C.T. Elvey Building, former location of ARSC

The Arctic Region Supercomputing Center (ARSC) was from 1993 to 2015 a research facility organized under the University of Alaska Fairbanks (UAF). Located on the UAF campus, ARSC offered high-performance computing (HPC) and mass storage to the UAF and State of Alaska research communities.

Contents

In general, the research supported with ARSC resources focused on the Earth's arctic region. Common projects included arctic weather modeling, Alaskan summer smoke forecasting, arctic sea ice analysis and tracking, Arctic Ocean systems, volcanic ash plume prediction, and tsunami forecasting and modeling.

ARSC was a Distributed Center (DC), an Allocated Distributed Center (ADC) and then one of six DoD Supercomputing Resource Centers (DSRCs) of the Department of Defense (DoD) High Performance Computing Modernization Program (HPCMP) from 1993 through 2011.

History

ARSC hosted a variety of HPC systems many of which were listed as among the Top 500 most powerful in the world. For more than 10 years ARSC maintained the standing of at least one system on the Top 500 list. [1] Funding for ARSC operations was primarily supplied by the DoD HPCMP, with augmentation through UAF and external grants and contracts from various sources such as the National Science Foundation. In December 2010, the Fairbanks Daily News-Miner reported probable layoffs for most of Arctic Region Supercomputer Center's 46 employees with the loss of its Department of Defense contract in 2011. [2] The article reported that 95 percent of ARSC funding comes from the Department of Defense. When that DoD funding source was lost ARSC could no longer afford computers that could be listed on the Top 500 List.

The following timeline includes various HPC systems acquired by ARSC and a Top 500 list standing when appropriate:

Related Research Articles

<span class="mw-page-title-main">Supercomputer</span> Type of extremely powerful computer

A supercomputer is a type of computer with a high level of performance as compared to a general-purpose computer. The performance of a supercomputer is commonly measured in floating-point operations per second (FLOPS) instead of million instructions per second (MIPS). Since 2017, supercomputers have existed, which can perform over 1017 FLOPS (a hundred quadrillion FLOPS, 100 petaFLOPS or 100 PFLOPS). For comparison, a desktop computer has performance in the range of hundreds of gigaFLOPS (1011) to tens of teraFLOPS (1013). Since November 2017, all of the world's fastest 500 supercomputers run on Linux-based operating systems. Additional research is being conducted in the United States, the European Union, Taiwan, Japan, and China to build faster, more powerful and technologically superior exascale supercomputers.

Floating point operations per second is a measure of computer performance in computing, useful in fields of scientific computations that require floating-point calculations.

<span class="mw-page-title-main">ETA10</span> 1980s supercomputer

The ETA10 is a vector supercomputer designed, manufactured, and marketed by ETA Systems, a spin-off division of Control Data Corporation (CDC). The ETA10 was an evolution of the CDC Cyber 205, which can trace its origins back to the CDC STAR-100, one of the first vector supercomputers to be developed.

<span class="mw-page-title-main">High-performance computing</span> Computing with supercomputers and clusters

High-performance computing (HPC) uses supercomputers and computer clusters to solve advanced computation problems.

<span class="mw-page-title-main">NEC SX</span>

NEC SX describes a series of vector supercomputers designed, manufactured, and marketed by NEC. This computer series is notable for providing the first computer to exceed 1 gigaflop, as well as the fastest supercomputer in the world between 1992–1993, and 2002–2004. The current model, as of 2018, is the SX-Aurora TSUBASA.

<span class="mw-page-title-main">Cray T3E</span>

The Cray T3E was Cray Research's second-generation massively parallel supercomputer architecture, launched in late November 1995. The first T3E was installed at the Pittsburgh Supercomputing Center in 1996. Like the previous Cray T3D, it was a fully distributed memory machine using a 3D torus topology interconnection network. The T3E initially used the DEC Alpha 21164 (EV5) microprocessor and was designed to scale from 8 to 2,176 Processing Elements (PEs). Each PE had between 64 MB and 2 GB of DRAM and a 6-way interconnect router with a payload bandwidth of 480 MB/s in each direction. Unlike many other MPP systems, including the T3D, the T3E was fully self-hosted and ran the UNICOS/mk distributed operating system with a GigaRing I/O subsystem integrated into the torus for network, disk and tape I/O.

<span class="mw-page-title-main">IBM Scalable POWERparallel</span> Series of supercomputers by IBM

Scalable POWERparallel (SP) is a series of supercomputers from IBM. SP systems were part of the IBM RISC System/6000 (RS/6000) family, and were also called the RS/6000 SP. The first model, the SP1, was introduced in February 1993, and new models were introduced throughout the 1990s until the RS/6000 was succeeded by eServer pSeries in October 2000. The SP is a distributed memory system, consisting of multiple RS/6000-based nodes interconnected by an IBM-proprietary switch called the High Performance Switch (HPS). The nodes are clustered using software called PSSP, which is mainly written in Perl.

<span class="mw-page-title-main">Roadrunner (supercomputer)</span> Former supercomputer built by IBM

Roadrunner was a supercomputer built by IBM for the Los Alamos National Laboratory in New Mexico, USA. The US$100-million Roadrunner was designed for a peak performance of 1.7 petaflops. It achieved 1.026 petaflops on May 25, 2008, to become the world's first TOP500 LINPACK sustained 1.0 petaflops system.

<span class="mw-page-title-main">TOP500</span> Database project devoted to the ranking of computers

The TOP500 project ranks and details the 500 most powerful non-distributed computer systems in the world. The project was started in 1993 and publishes an updated list of the supercomputers twice a year. The first of these updates always coincides with the International Supercomputing Conference in June, and the second is presented at the ACM/IEEE Supercomputing Conference in November. The project aims to provide a reliable basis for tracking and detecting trends in high-performance computing and bases rankings on HPL benchmarks, a portable implementation of the high-performance LINPACK benchmark written in Fortran for distributed-memory computers.

The United States Department of Defense High Performance Computing Modernization Program (HPCMP) was initiated in 1992 in response to Congressional direction to modernize the Department of Defense (DoD) laboratories’ high performance computing capabilities. The HPCMP provides supercomputers, a national research network, high-end software tools, a secure environment, and computational science experts that together enable the Defense laboratories and test centers to conduct research, development, test and technology evaluation activities.

<span class="mw-page-title-main">Magerit</span> Supercomputer in Madrid, Spain

Magerit is one of the most powerful supercomputers in Spain. It also reached the second best Spanish position in the TOP500 list of supercomputers. It is installed in CeSViMa, a research center of the Technical University of Madrid.

<span class="mw-page-title-main">Jaguar (supercomputer)</span> Cray supercomputer at Oak Ridge National Laboratory

Jaguar or OLCF-2 was a petascale supercomputer built by Cray at Oak Ridge National Laboratory (ORNL) in Oak Ridge, Tennessee. The massively parallel Jaguar had a peak performance of just over 1,750 teraFLOPS. It had 224,256 x86-based AMD Opteron processor cores, and operated with a version of Linux called the Cray Linux Environment. Jaguar was a Cray XT5 system, a development from the Cray XT4 supercomputer.

This list compares various amounts of computing power in instructions per second organized by order of magnitude in FLOPS.

The Swiss National Supercomputing Centre is the national high-performance computing centre of Switzerland. It was founded in Manno, canton Ticino, in 1991. In March 2012, the CSCS moved to its new location in Lugano-Cornaredo.

<span class="mw-page-title-main">K computer</span> Supercomputer in Kobe, Japan

The K computer – named for the Japanese word/numeral "kei" (京), meaning 10 quadrillion (1016) – was a supercomputer manufactured by Fujitsu, installed at the Riken Advanced Institute for Computational Science campus in Kobe, Hyōgo Prefecture, Japan. The K computer was based on a distributed memory architecture with over 80,000 compute nodes. It was used for a variety of applications, including climate research, disaster prevention and medical research. The K computer's operating system was based on the Linux kernel, with additional drivers designed to make use of the computer's hardware.

<span class="mw-page-title-main">History of supercomputing</span>

The history of supercomputing goes back to the 1960s when a series of computers at Control Data Corporation (CDC) were designed by Seymour Cray to use innovative designs and parallelism to achieve superior computational peak performance. The CDC 6600, released in 1964, is generally considered the first supercomputer. However, some earlier computers were considered supercomputers for their day such as the 1954 IBM NORC in the 1950s, and in the early 1960s, the UNIVAC LARC (1960), the IBM 7030 Stretch (1962), and the Manchester Atlas (1962), all of which were of comparable power.

<span class="mw-page-title-main">Supercomputing in Europe</span> Overview of supercomputing in Europe

Several centers for supercomputing exist across Europe, and distributed access to them is coordinated by European initiatives to facilitate high-performance computing. One such initiative, the HPC Europa project, fits within the Distributed European Infrastructure for Supercomputing Applications (DEISA), which was formed in 2002 as a consortium of eleven supercomputing centers from seven European countries. Operating within the CORDIS framework, HPC Europa aims to provide access to supercomputers across Europe.

XK7 is a supercomputing platform, produced by Cray, launched on October 29, 2012. XK7 is the second platform from Cray to use a combination of central processing units ("CPUs") and graphical processing units ("GPUs") for computing; the hybrid architecture requires a different approach to programming to that of CPU-only supercomputers. Laboratories that host XK7 machines host workshops to train researchers in the new programming languages needed for XK7 machines. The platform is used in Titan, the world's second fastest supercomputer in the November 2013 list as ranked by the TOP500 organization. Other customers include the Swiss National Supercomputing Centre which has a 272 node machine and Blue Waters has a machine that has Cray XE6 and XK7 nodes that performs at approximately 1 petaFLOPS (1015 floating-point operations per second).

The Cray XC30 is a massively parallel multiprocessor supercomputer manufactured by Cray. It consists of Intel Xeon processors, with optional Nvidia Tesla or Xeon Phi accelerators, connected together by Cray's proprietary "Aries" interconnect, stored in air-cooled or liquid-cooled cabinets. Each liquid-cooled cabinet can contain up to 48 blades, each with eight CPU sockets, and uses 90 kW of power. The XC series supercomputers are available with the Cray DataWarp applications I/O accelerator technology.

<span class="mw-page-title-main">Aurora (supercomputer)</span> Planned supercomputer

Aurora is a supercomputer that was sponsored by the United States Department of Energy (DOE) and designed by Intel and Cray for the Argonne National Laboratory. It has been the second fastest supercomputer in the world since 2023. It is expected that after optimizing its performance it will exceed 2 ExaFLOPS, making it the fastest computer ever.

References

  1. For more than 10 years ARSC maintained the standing of at least one system on the Top 500 list. "University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2010-06-30. Retrieved 2019-07-06.
  2. ARSC, located at the University of Alaska Fairbanks since 1993, had a budget of about $12 million last year. About 95 percent of that funding comes from the Department of Defense, which has a local contract to provide military and security assistance.Jeff Richardson (2010-12-03). "Arctic Region Supercomputer Center may lose DoD contract; layoffs expected". newsminer.com. Fairbanks Daily News-Miner. Retrieved 2019-07-06.
  3. With the Cray Y-MP named Denali Arctic Region Supercomputing Center (ARSC) was #251 on the first Top 500 Supercompter list in June 1993 published on the first day of the 8th Mannheim Supercomputer Seminar. "TOP500 List - June 1993". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1993-06-30. Retrieved 2019-07-06.
  4. This Cray Y-MP M98/41024 system was #251 on the 06/1993 Top 500 list, #302 on the 11/1993 Top 500 list, #405 on the 06/1994 Top 500 list. "Cray Y-MP M98/41024 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1994-06-30. Retrieved 2019-07-06.
  5. A Cray T3D named Yukon got #58 on the June 1994 Top 500 Supercompter list while the previous computer Denali was still on the list at position #405 giving ARSC two of the Top 500 Supercompters at the same time. "TOP500 List - June 1994". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1994-06-30. Retrieved 2019-07-06.
  6. This ARSC Cray T3D MC128-2 was #58 on the June 1994 Top 500 list and #55 on the November 1994 Top 500 list. "T3D MC128-2 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1994-06-30. Retrieved 2019-07-06.
  7. The Cray T3D named Yukon was upgraded to a Cray T3D MC128-8 for spot #83 on the June 1995 Top 500 list. "TOP500 List - June 1995". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1994-06-30. Retrieved 2019-07-06.
  8. This ARSC Cray T3D MC128-8 was on the Top 500 list for three years: #83 06/1995, #99 12/1995, #127 06/1996, #169 11/1996, #239 06/1997 and #340 11/1997 "Cray T3D MC128-8 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-11-30. Retrieved 2019-07-06.
  9. The Cray T3D named Yukon was upgraded again to a Cray T3E got get ARSC position #70 on the June 1997 Top 500 list. "TOP500 List - June 1997". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-06-30. Retrieved 2019-07-06.
  10. Yukon was only on the Top 500 list once in this configuration. "Cray T3E at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-06-30. Retrieved 2019-07-06.
  11. Another upgrade to the T3E900 got ARSC position #62 on the November 1997 Top 500 list. "TOP500 List - November 1997". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-11-30. Retrieved 2019-07-06.
  12. Yukon was on the Top 500 list in this configuration as #62 in 11/1997 and as #67 in 06/1998. "Cray T3E900 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-06-30. Retrieved 2019-07-06.
  13. "Arctic Region Supercomputing Center installs Cray T3E". HPC Wire. Tabor Communications. 1997-03-07. Retrieved 2019-07-06.
  14. ARSC gets #74 ARSC gets #74 on the November 1998 Top 500 list with another upgrade to the T3E900. "TOP500 List - November 1998". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1998-11-30. Retrieved 2019-07-06.
  15. Yukon was only on the Top 500 list once in this configuration. "Cray T3E900 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1997-06-30. Retrieved 2019-07-06.
  16. ARSC got spot #44 on the June 1999 Top 500 list after Yukon was upgraded again to a Cray T3E900 with 268 cores. "TOP500 List - June 1999". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 1999-06-30. Retrieved 2019-07-06.
  17. Yukon is on the Top 500 list in this configuration for over three years: #44 06/1999, #56 11/1999, #78 06/2000, #107 11/2000, #131 06/2001, #199 11/2001 and #383 06/2002 "Cray T3E900 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2002-06-30. Retrieved 2019-07-06.
  18. The IBM Icehawk RISC system got ARSC on the Top 500 list as #117 in June of 2001, #147 in November of 2001, #264 in June of 2002 and #367 in November of 2002. "IBM SP Power3 375 MHz at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2002-11-30. Retrieved 2019-07-06.
  19. An IBM SP named Icehawk with 200 CPUs and 276 GFLOPS got ARSC the #117 spot on the June 2001 Top 500 list. "TOP500 List - June 2001". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2001-06-30. Retrieved 2019-07-06.
  20. Stephen Shankland (2002-01-02). "Compaq, Cray: Supercomputer progress". CNET tells you what's new in tech, culture and science. CBS Interactive Inc. Retrieved 2019-07-06.
  21. ARSC got spot #116 on the June 2003 Top 500 list with a Cray X1 named Klondike with 60 cores. "TOP500 List - June 2003". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2003-06-30. Retrieved 2019-07-06.
  22. Klondike was only on the Top 500 list once in this configuration. "Cray X1 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2003-06-30. Retrieved 2019-07-06.
  23. Upgrading the Cray X1 to 124 cores got ARSC spot #71 on the November 2003 Top 500 list. "TOP500 List - November 2003". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2003-11-30. Retrieved 2019-07-06.
  24. Klondike in this configuration on the Top 500 list as #71 in 11/2003, #154 in 06/2004, #202 in 11/2004 and #353 in 06/2005. "Cray X1 at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2003-06-30. Retrieved 2019-07-06.
  25. Stephen Shankland (2003-03-13). "IBM nabs hot supercomputer deal in arctic". CNET tells you what's new in tech, culture and science. CBS Interactive Inc. Retrieved 2019-07-06.
  26. ARSC got spot #56 on the June 2004 Top 500 list with its IBM System p named Iceberg. "TOP500 List - June 2004". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2004-06-30. Retrieved 2019-07-06.
  27. In 2003 IBM had told InfoWorld that Iceberg would put ARSC in sixth place on the Top 500 list.Tom Krazit (2003-03-13). "IBM supercomputer heads north - Arctic center to use 800-processor box". InfoWorld a leading voice in emerging enterprise technology. IDG Communications, Inc. Retrieved 2019-07-06.
  28. The IBM Iceberg RISC system got ARSC Top 500 list spots #56 in 06/2004, #78 in 11/2004, #136 in 06/2005, #253 in 11/2005 and #398 in 06/2006. "IBM eServer pSeries 655 (1.5 GHz Power4+) at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2006-06-30. Retrieved 2019-07-06.
  29. Midnight was only on the Top 500 list once. "Midnight - Fire x2200/x4600 Cluster, Opteron 2.6 GHz, Infiniband, Linux system at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2007-11-30. Retrieved 2019-07-06.
  30. A Sun Opteron Cluster named Midnight with 2,236 cores and 12 TFLOPS got ARSC the #206 spot on the November 2007 Top 500 list. "TOP500 List - November 2007". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2007-11-30. Retrieved 2019-07-06.
  31. Pingo got ARSC Top 500 list spots #109 in 11/2008, #205 in 06/2009, #290 in 11/2009 and #435 in 06/2010. "Cray XT5 QC 2.3 GHz at University of Alaska - Arctic Region Supercomputing Center". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2010-06-30. Retrieved 2019-07-06.
  32. A Cray XT5 name Pingo with 3,456 cores got ARSC the #109 spot on the November 2008 Top 500 list. "TOP500 List - November 2008". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2008-11-30. Retrieved 2019-07-06.
  33. A pre-Thanksgiving email to ARSC confirmed what many at the center had suspected, namely that the center would lose its DoD funding after the current money expires next May. Today the center is funded to the tune of $12 to $15 million, and the DoD slice represents around 95 percent of the total.Michael Feldman (2010-01-22). "Arctic Region Supercomputing Center Gets Cold Shoulder from DoD". HPC Wire. Tabor Communications. Retrieved 2019-07-06.
  34. "Cray XE6, Opteron 16C 2.500GHz, Cray Gemini interconnect". The TOP500 List of the 500 most powerful commercially available computer systems known. The TOP500 project. 2013-06-30. Retrieved 2019-07-06.
  35. "Penguin Computing Supplies PACMAN Supercomputer to University of Alaska Fairbanks". Penguin Computing. SMART Global Holdings. 2011-05-24. Retrieved 2019-07-06.
  36. "Arctic Region Supercomputing Center - Chugach Cray XE6 system". Internet Archive Wayback Machine. The Internet Archive, a 501(c)(3) non-profit. Archived from the original on 2011-10-09. Retrieved 2019-07-07.
  37. "Arctic Region Supercomputing Center Staff". Internet Archive Wayback Machine. The Internet Archive, a 501(c)(3) non-profit. Archived from the original on 2012-01-07. Retrieved 2019-07-07.
  38. "Arctic Region Supercomputing Center Staff". Internet Archive Wayback Machine. The Internet Archive, a 501(c)(3) non-profit. Archived from the original on 2014-12-22. Retrieved 2019-07-07.
  39. "Arctic Region Supercomputing Center". Internet Archive Wayback Machine. The Internet Archive, a 501(c)(3) non-profit. Archived from the original on 2011-11-02. Retrieved 2019-07-07.

64°51′36″N147°50′57″W / 64.8600°N 147.8491°W / 64.8600; -147.8491