TORQUE

Last updated
TORQUE
Developer(s) Adaptive computing
Initial release2003 (2003)
Stable release
7.0 / 27 January 2023;14 months ago (2023-01-27)
Written in ANSI C
Operating system Unix-like
Size 5 MB
Available in English
Type Distributed resource manager
License Proprietary License (As of June 2018), [1] OpenPBS version 2.3 [2] [3] (non-free in DFSG [4] )
Website adaptivecomputing.com/cherry-services/torque-resource-manager/

The Terascale Open-source Resource and Queue Manager (TORQUE) is a distributed resource manager designed to oversee batch jobs and distributed compute nodes. [5] It offers control and management capabilities for clusters, aiding in utilization, scheduling, and administration tasks.

Contents

TORQUE can be integrated with either the non-commercial Maui Cluster Scheduler or the commercial Moab Workload Manager, providing enhanced functionality and optimization for cluster environments.

Initially based on the Portable Batch System (PBS), the TORQUE community has expanded its capabilities to improve scalability, fault tolerance, and overall functionality. Notable contributors to TORQUE include organizations such as NCSA, OSC, USC, the US DOE, Sandia, PNNL, UB, TeraGrid and other High-Performance Computing (HPC) entities.

As of June 2018, TORQUE is no longer considered open-source software due to licensing issues. While it was previously described as open-source software and utilized the OpenPBS version 2.3 license, it is now categorized as non-free software according to the Debian Free Software Guidelines. [1] [2] [4]

See also

Related Research Articles

Computerized batch processing is a method of running software programs called jobs in batches automatically. While users are required to submit the jobs, no other interaction by the user is required to process the batch. Batches may automatically be run at scheduled times as well as being run contingent on the availability of computer resources.

The Debian Free Software Guidelines (DFSG) is a set of guidelines that the Debian Project uses to determine whether a software license is a free software license, which in turn is used to determine whether a piece of software can be included in Debian. The DFSG is part of the Debian Social Contract.

MOSIX is a proprietary distributed operating system. Although early versions were based on older UNIX systems, since 1999 it focuses on Linux clusters and grids. In a MOSIX cluster/grid there is no need to modify or to link applications with any library, to copy files or login to remote nodes, or even to assign processes to different nodes – it is all done automatically, like in an SMP.

Oracle Grid Engine, previously known as Sun Grid Engine (SGE), CODINE or GRD, was a grid computing computer cluster software system, acquired as part of a purchase of Gridware, then improved and supported by Sun Microsystems and later Oracle. There have been open source versions and multiple commercial versions of this technology, initially from Sun, later from Oracle and then from Univa Corporation.

HTCondor is an open-source high-throughput computing software framework for coarse-grained distributed parallelization of computationally intensive tasks. It can be used to manage workload on a dedicated cluster of computers, or to farm out work to idle desktop computers – so-called cycle scavenging. HTCondor runs on Linux, Unix, Mac OS X, FreeBSD, and Microsoft Windows operating systems. HTCondor can integrate both dedicated resources and non-dedicated desktop machines into one computing environment.

A job scheduler is a computer application for controlling unattended background program execution of jobs. This is commonly called batch scheduling, as execution of non-interactive jobs is often called batch processing, though traditional job and batch are distinguished and contrasted; see that page for details. Other synonyms include batch system, distributed resource management system (DRMS), distributed resource manager (DRM), and, commonly today, workload automation (WLA). The data structure of jobs to run is known as the job queue.

Maui Cluster Scheduler is a job scheduler for use on clusters and supercomputers initially developed by Cluster Resources, Inc. Maui is capable of supporting multiple scheduling policies, dynamic priorities, reservations, and fairshare capabilities.

Portable Batch System is the name of computer software that performs job scheduling. Its primary task is to allocate computational tasks, i.e., batch jobs, among the available computing resources. It is often used in conjunction with UNIX cluster environments.

<span class="mw-page-title-main">Platform Computing</span> Software company in Canada

Platform Computing was a privately held software company primarily known for its job scheduling product, Load Sharing Facility (LSF). It was founded in 1992 in Toronto, Ontario, Canada and headquartered in Markham, Ontario with 11 branch offices across the United States, Europe and Asia.

Open Source Cluster Application Resources (OSCAR) is a Linux-based software installation for high-performance cluster computing. OSCAR allows users to install a Beowulf type high performance computing cluster.

<span class="mw-page-title-main">Univa</span> Software company

Univa was a software company that developed workload management and cloud management products for compute-intensive applications in the data center and across public, private, and hybrid clouds, before being acquired by Altair Engineering in September 2020.

Windows HPC Server 2008, released by Microsoft on 22 September 2008, is the successor product to Windows Compute Cluster Server 2003. Like WCCS, Windows HPC Server 2008 is designed for high-end applications that require high performance computing clusters. This version of the server software is claimed to efficiently scale to thousands of cores. It includes features unique to HPC workloads: a new high-speed NetworkDirect RDMA, highly efficient and scalable cluster management tools, a service-oriented architecture (SOA) job scheduler, an MPI library based on open-source MPICH2, and cluster interoperability through standards such as the High Performance Computing Basic Profile (HPCBP) specification produced by the Open Grid Forum (OGF).

gLite Grid computing software

gLite is a middleware computer software project for grid computing used by the CERN LHC experiments and other scientific domains. It was implemented by collaborative efforts of more than 80 people in 12 different academic and industrial research centers in Europe. gLite provides a framework for building applications tapping into distributed computing and storage resources across the Internet. The gLite services were adopted by more than 250 computing centres, and used by more than 15000 researchers in Europe and around the world.

<span class="mw-page-title-main">Slurm Workload Manager</span> Free and open-source job scheduler for Linux and similar computers

The Slurm Workload Manager, formerly known as Simple Linux Utility for Resource Management (SLURM), or simply Slurm, is a free and open-source job scheduler for Linux and Unix-like kernels, used by many of the world's supercomputers and computer clusters.

Univa Grid Engine (UGE) is a batch-queuing system, forked from Sun Grid Engine (SGE). The software schedules resources in a data center applying user-configurable policies to help improve resource sharing and throughput by maximizing resource utilization. The product can be deployed to run on-premises, using IaaS cloud computing or in a hybrid cloud environment.

<span class="mw-page-title-main">DIET</span>

DIET is a software for grid-computing. As middleware, DIET sits between the operating system and the application software. DIET was created in 2000. It was designed for high-performance computing. It is currently developed by INRIA, École Normale Supérieure de Lyon, CNRS, Claude Bernard University Lyon 1, SysFera. It is open-source software released under the CeCILL license.

Carter is a supercomputer installed at Purdue University in the fall of 2011 in a partnership with Intel. The high-performance computing cluster is operated by Information Technology at Purdue (ITaP), the university's central information technology organization. ITaP also operates clusters named Steele built in 2007, Coates built in 2009, Rossmann built in 2010, and Hansen built in the summer of 2011. Carter was the fastest campus supercomputer in the U.S. outside a national center when built. It was one of the first clusters to employ Intel's second generation Xenon E-5 "Sandy Bridge" processor and ranked 54th on the November 2011 TOP500 list, making it Purdue's first Top 100-ranked research computing system.

Bright Computing, Inc. is a developer of software for deploying and managing high-performance (HPC) clusters, Kubernetes clusters, and OpenStack private clouds in on-premises data centers as well as in the public cloud.

<span class="mw-page-title-main">Singularity (software)</span> Free, cross-platform and open-source computer program

Singularity is a free and open-source computer program that performs operating-system-level virtualization also known as containerization.

References

  1. 1 2 "Closed Source Software License". Adaptive Computing, Inc. 2018. Retrieved 2018-07-31.
  2. 1 2 Veridian Information Solutions, Inc. (2000). "OpenPBS (Portable Batch System) v2.3 Software License". Cluster Resources, Inc. Archived from the original on 2011-08-20. Retrieved 2011-07-31.
  3. "Torque resource manager". Cluster Resources, Inc. 2011. Archived from the original on 2011-07-19. Retrieved 2011-07-31.
  4. 1 2 "The DFSG and Software Licenses - Licenses that are DFSG-incompatible". Debian. 2011-03-27. Archived from the original on 2011-07-25. Retrieved 2011-07-31.
  5. TORQUE resource manager, Garrick Staples, SC '06: Proceedings of the 2006 ACM/IEEE conference on Supercomputing, ISBN   0-7695-2700-0