Yellowstone (supercomputer)

Last updated • 2 min readFrom Wikipedia, The Free Encyclopedia
Yellowstone
NSF-supported Research Facilities (15947776872).jpg
Active2012-2017
Architecturex86 (iDataPlex)
Memory145 terabytes
Storage320 petabytes (storage)
36.4 petabytes (online)
Speed1.5 petaflops

Yellowstone [1] was the inaugural supercomputer at the NCAR-Wyoming Supercomputing Center [2] (NWSC) in Cheyenne, Wyoming. It was installed, tested, and readied for production in the summer of 2012. [3] The Yellowstone supercomputing cluster was decommissioned on December 31, 2017, [4] being replaced by its successor Cheyenne. [5]

Contents

Yellowstone was a highly capable petascale system designed for conducting breakthrough scientific research in the interdisciplinary field of Earth system science. Scientists used the computer and its associated resources to model and analyze complex processes in the atmosphere, oceans, ice caps, and throughout the Earth system, accelerating scientific research in climate change, severe weather, geomagnetic storms, carbon sequestration, aviation safety, wildfires, and many other topics. [6] [7] Funded by the National Science Foundation and the State and University of Wyoming, and operated by the National Center for Atmospheric Research, Yellowstone's purpose was to improve the predictive power of Earth system science simulation to benefit decision-making and planning for society. [8]

System description

Yellowstone was a 1.5-petaflops IBM iDataPlex cluster computer with 4,536 dual-socket compute nodes that contained 9,072, 2.6-GHz Intel Xeon E5-2670 8-core processors (72,576 cores), and its aggregate memory size was 145 terabytes. [9] The nodes interconnected in a full fat tree network via a Mellanox FDR InfiniBand switching fabric. [9] System software [10] includes the Red Hat Enterprise Linux operating system for Scientific Computing, [11] LSF Batch Subsystem and Resource Manager, [12] and IBM General Parallel File System (GPFS). [9]

Yellowstone was integrated with many other high-performance computing resources in the NWSC. The central feature of this supercomputing architecture was its shared file system that streamlined science workflows by providing computation, analysis, and visualization work spaces common to all resources. This common data storage pool, called the GLobally Accessible Data Environment [13] (GLADE), provides 36.4 petabytes of online disk capacity shared by the supercomputer, two data analysis and visualization (DAV) cluster computers (Geyser and Caldera), [9] data servers for both local and remote users, and a data archive with the capacity to store 320 petabytes of research data. High-speed networks connect this Yellowstone environment to science gateways, [14] data transfer services, remote visualization resources, Extreme Science and Engineering Discovery Environment (XSEDE) sites, and partner sites around the world.

This integration of computing resources, file systems, data storage, and broadband networks allowed scientists to simulate future geophysical scenarios at high resolution, then analyze and visualize them on one computing complex. [15] This improves scientific productivity [6] by avoiding the delays associated with moving large quantities of data between separate systems. Further, this reduces the volume of data that needs to be transferred to researchers at their home institutions. The Yellowstone environment at NWSC makes more than 600 million processor-hours available each year to researchers in the Earth system sciences. [16]

See also

Related Research Articles

<span class="mw-page-title-main">Supercomputer</span> Type of extremely powerful computer

A supercomputer is a type of computer with a high level of performance as compared to a general-purpose computer. The performance of a supercomputer is commonly measured in floating-point operations per second (FLOPS) instead of million instructions per second (MIPS). Since 2017, supercomputers have existed which can perform over 1017 FLOPS (a hundred quadrillion FLOPS, 100 petaFLOPS or 100 PFLOPS). For comparison, a desktop computer has performance in the range of hundreds of gigaFLOPS (1011) to tens of teraFLOPS (1013). Since November 2017, all of the world's fastest 500 supercomputers run on Linux-based operating systems. Additional research is being conducted in the United States, the European Union, Taiwan, Japan, and China to build faster, more powerful and technologically superior exascale supercomputers.

<span class="mw-page-title-main">National Center for Supercomputing Applications</span> Illinois-based applied supercomputing research organization

The National Center for Supercomputing Applications (NCSA) is a state-federal partnership to develop and deploy national-scale cyberinfrastructure that advances research, science and engineering based in the United States. NCSA operates as a unit of the University of Illinois Urbana-Champaign, and provides high-performance computing resources to researchers across the country. Support for NCSA comes from the National Science Foundation, the state of Illinois, the University of Illinois, business and industry partners, and other federal agencies.

<span class="mw-page-title-main">IBM Blue Gene</span> Series of supercomputers by IBM

Blue Gene was an IBM project aimed at designing supercomputers that can reach operating speeds in the petaFLOPS (PFLOPS) range, with low power consumption.

<span class="mw-page-title-main">MareNostrum</span> Supercomputer in the Barcelona Supercomputing Center

MareNostrum is the main supercomputer in the Barcelona Supercomputing Center. It is the most powerful supercomputer in Spain, one of thirteen supercomputers in the Spanish Supercomputing Network and one of the seven supercomputers of the European infrastructure PRACE.

<span class="mw-page-title-main">NASA Advanced Supercomputing Division</span> Provides computing resources for various NASA projects

The NASA Advanced Supercomputing (NAS) Division is located at NASA Ames Research Center, Moffett Field in the heart of Silicon Valley in Mountain View, California. It has been the major supercomputing and modeling and simulation resource for NASA missions in aerodynamics, space exploration, studies in weather patterns and ocean currents, and space shuttle and aircraft design and development for almost forty years.

<span class="mw-page-title-main">TeraGrid</span>

TeraGrid was an e-Science grid computing infrastructure combining resources at eleven partner sites. The project started in 2001 and operated from 2004 through 2011.

<span class="mw-page-title-main">National Center for Atmospheric Research</span> US federally funded research and development center

The US National Center for Atmospheric Research is a US federally funded research and development center (FFRDC) managed by the nonprofit University Corporation for Atmospheric Research (UCAR) and funded by the National Science Foundation (NSF). NCAR has multiple facilities, including the I. M. Pei-designed Mesa Laboratory headquarters in Boulder, Colorado. Studies include meteorology, climate science, atmospheric chemistry, solar-terrestrial interactions, environmental and societal impacts.

The Texas Advanced Computing Center (TACC) at the University of Texas at Austin, United States, is an advanced computing research center that is based on comprehensive advanced computing resources and supports services to researchers in Texas and across the U.S. The mission of TACC is to enable discoveries that advance science and society through the application of advanced computing technologies. Specializing in high-performance computing, scientific visualization, data analysis & storage systems, software, research & development, and portal interfaces, TACC deploys and operates advanced computational infrastructure to enable the research activities of faculty, staff, and students of UT Austin. TACC also provides consulting, technical documentation, and training to support researchers who use these resources. TACC staff members conduct research and development in applications and algorithms, computing systems design/architecture, and programming tools and environments.

<span class="mw-page-title-main">National Energy Research Scientific Computing Center</span> Supercomputer facility operated by the US Department of Energy in Berkeley, California

The National Energy Research Scientific Computing Center (NERSC), is high-performance computing (supercomputer) .National User Facility is operated by Lawrence Berkeley National Laboratory for the United States Department of Energy Office of Science. The mission is to establish a computing center for the Office of Science, NERSC houses high performance computing and data systems which can be used by 9,000 scientists at national laboratories and universities around the country. Research at NERSC is focused on fundamental and applied research with energy efficiency, storage, generation and Earth systems science, understanding of fundamental forces of nature and the Universe. The largest research areas are High Energy Physics, Materials Science, Chemical Sciences, Climate and Environmental Sciences, Nuclear Physics, and Fusion Energy research. NERSC's newest and largest supercomputer is Perlmutter, which debuted in 2021 and ranked 5th on the TOP500 list of world's fastest supercomputers.

<span class="mw-page-title-main">Roadrunner (supercomputer)</span> Former supercomputer built by IBM

Roadrunner was a supercomputer built by IBM for the Los Alamos National Laboratory in New Mexico, USA. The US$100-million Roadrunner was designed for a peak performance of 1.7 petaflops. It achieved 1.026 petaflops on May 25, 2008, to become the world's first TOP500 LINPACK sustained 1.0 petaflops system.

<span class="mw-page-title-main">TOP500</span> Database project devoted to the ranking of computers

The TOP500 project ranks and details the 500 most powerful non-distributed computer systems in the world. The project was started in 1993 and publishes an updated list of the supercomputers twice a year. The first of these updates always coincides with the International Supercomputing Conference in June, and the second is presented at the ACM/IEEE Supercomputing Conference in November. The project aims to provide a reliable basis for tracking and detecting trends in high-performance computing and bases rankings on HPL benchmarks, a portable implementation of the high-performance LINPACK benchmark written in Fortran for distributed-memory computers.

<span class="mw-page-title-main">Supercomputing and Visualization Center of Madrid</span> Supercomputer site in Madrid, Spain

The Supercomputing and Visualization Center of Madrid (CeSViMa), also called Madrid Supercomputing and Visualization Center, depends on the computer science faculty of the Technical University of Madrid. This center houses Magerit, one of the most powerful supercomputers in Spain. This center is a member of the Spanish Supercomputing Network, the Spanish e-Science Network and the Madrid Laboratories and Infraestructures Network.

<span class="mw-page-title-main">Pleiades (supercomputer)</span> NASA supercomputer at Ames Research Center/NAS

Pleiades is a petascale supercomputer housed at the NASA Advanced Supercomputing (NAS) facility at NASA's Ames Research Center located at Moffett Field near Mountain View, California. It is maintained by NASA and partners Hewlett Packard Enterprise and Intel.

The National Center for Computational Sciences (NCCS) is a United States Department of Energy (DOE) Leadership Computing Facility that houses the Oak Ridge Leadership Computing Facility (OLCF), a DOE Office of Science User Facility charged with helping researchers solve challenging scientific problems of global interest with a combination of leading high-performance computing (HPC) resources and international expertise in scientific computing.

SuperMUC was a supercomputer of the Leibniz Supercomputing Centre (LRZ) of the Bavarian Academy of Sciences. It was housed in the LRZ's data centre in Garching near Munich. It was decommissioned in January 2020, having been superseded by the more powerful SuperMUC-NG.

<span class="mw-page-title-main">Supercomputing in Europe</span> Overview of supercomputing in Europe

Several centers for supercomputing exist across Europe, and distributed access to them is coordinated by European initiatives to facilitate high-performance computing. One such initiative, the HPC Europa project, fits within the Distributed European Infrastructure for Supercomputing Applications (DEISA), which was formed in 2002 as a consortium of eleven supercomputing centers from seven European countries. Operating within the CORDIS framework, HPC Europa aims to provide access to supercomputers across Europe.

<span class="mw-page-title-main">Mira (supercomputer)</span>

Mira is a retired petascale Blue Gene/Q supercomputer. As of November 2017, it is listed on TOP500 as the 11th fastest supercomputer in the world, while it debuted June 2012 in 3rd place. It has a performance of 8.59 petaflops (LINPACK) and consumes 3.9 MW. The supercomputer was constructed by IBM for Argonne National Laboratory's Argonne Leadership Computing Facility with the support of the United States Department of Energy, and partially funded by the National Science Foundation. Mira was used for scientific research, including studies in the fields of material science, climatology, seismology, and computational chemistry. The supercomputer was used initially for sixteen projects selected by the Department of Energy.

<span class="mw-page-title-main">NCAR-Wyoming Supercomputing Center</span> High performance computing center in Wyoming, US

The NCAR-Wyoming Supercomputing Center (NWSC) is a high-performance computing (HPC) and data archival facility located in Cheyenne, Wyoming, that provides advanced computing services to researchers in the Earth system sciences.

<span class="mw-page-title-main">Cray XC40</span> Supercomputer manufactured by Cray

The Cray XC40 is a massively parallel multiprocessor supercomputer manufactured by Cray. It consists of Intel Haswell Xeon processors, with optional Nvidia Tesla or Intel Xeon Phi accelerators, connected together by Cray's proprietary "Aries" interconnect, stored in air-cooled or liquid-cooled cabinets. The XC series supercomputers are available with the Cray DataWarp applications I/O accelerator technology.

The Cheyenne supercomputer at the NCAR-Wyoming Supercomputing Center (NWSC) in Cheyenne, Wyoming operated for seven years as one of the world’s most powerful and energy-efficient computers from 2017 to 2024. Ranked in November 2016 as the 20th most powerful computer in the world and November 2023 as 160th by Top500, the 5.34-petaflops system is capable of more than triple the amount of scientific computing performed by NCAR’s previous supercomputer, Yellowstone. It also is three times more energy efficient than Yellowstone, with a peak computation rate of more than 3 billion calculations per second for every watt of energy consumed.

References

  1. "Yellowstone", NCAR Computational and Information Systems Laboratory (CISL) website: Resources. Retrieved 2012-06-12.
  2. "NCAR-Wyoming Supercomputing Center Fact Sheet", University Corporation for Atmospheric Research (UCAR) website, Retrieved 2012-06-12.
  3. NCAR Advances Weather Research Capabilities With IBM Supercomputing Technology, IBM News Release, 08 Nov 2011.
  4. "Yellowstone to be decommissioned December 31 | Computational & Information Systems Laboratory". dailyb.cisl.ucar.edu. Retrieved 2018-01-19.
  5. Scoles, Sarah (31 March 2017). "Why You Should Put Your Supercomputer in Wyoming". Wired.com. Wired. Retrieved 6 October 2018.
  6. 1 2 NCAR Selects IBM for Key Components of New Supercomputing Center, NCAR/UCAR AtmosNews, 7 November 2011.
  7. Yellowstone, NWSC science impact, NCAR Computational and Information Systems Laboratory (CISL) website: Resources. Retrieved 2012-06-12.
  8. The NCAR-Wyoming Supercomputing Center Science Justification, Proposal to The National Science Foundation by The National Center for Atmospheric Research and The University Corporation for Atmospheric Research in partnership with The University and State of Wyoming, 4 September 2009.
  9. 1 2 3 4 System overview, Yellowstone: High-performance computing resource, NCAR Computational and Information Systems Laboratory (CISL) website: Resources. Retrieved 2012-06-12.
  10. Yellowstone Software, NCAR Computational and Information Systems Laboratory (CISL) website: Resources. Retrieved 2012-06-12.
  11. Red Hat Enterprise Linux For Scientific Computing, Red Hat Products website, Retrieved 2012-06-12.
  12. [Note that IBM has acquired Platform Computing, Inc., developers of LSF.]
  13. NCAR’s Globally Accessible Data Environment, FY2011 CISL Annual Report. Note: This October 2011 report describes GLADE at NCAR’s Mesa Lab Computing Facility in Boulder, Colorado. The design of GLADE at NWSC in Cheyenne, Wyoming is identical at this level of description.
  14. Science gateway services, FY2011 CISL Annual Report.
  15. NCAR to Install 1.6 Petaflop IBM Supercomputer, HPCwire, November 07, 2011.
  16. NCAR's next supercomputer: Yellowstone, News@Unidata, 22 November 2011.