GROMACS

Last updated
GROMACS
Developer(s) University of Groningen
Royal Institute of Technology
Uppsala University [1]
Initial release1991;33 years ago (1991)
Stable release
2023.3 / 19 October 2023;6 months ago (2023-10-19) [2]
Repository
Written in C++, C, CUDA, OpenCL, SYCL
Operating system Linux, macOS, Windows, any other Unix variety
Platform Many
Available inEnglish
Type Molecular dynamics simulation
License LGPL versions >= 4.6,
GPL versions < 4.6 [3]
Website www.gromacs.org

GROMACS is a molecular dynamics package mainly designed for simulations of proteins, lipids, and nucleic acids. It was originally developed in the Biophysical Chemistry department of University of Groningen, and is now maintained by contributors in universities and research centers worldwide. [4] [5] [6] GROMACS is one of the fastest and most popular software packages available, [7] [8] and can run on central processing units (CPUs) and graphics processing units (GPUs). [9] It is free, open-source software released under the GNU General Public License (GPL), [3] and starting with version 4.6, the GNU Lesser General Public License (LGPL).

Contents

History

The GROMACS project originally began in 1991 at Department of Biophysical Chemistry, University of Groningen, Netherlands (1991–2000). Its name originally derived from this time (GROningen MAchine for Chemical Simulations) although currently GROMACS is not an abbreviation for anything, as little active development has taken place in Groningen in recent decades. The original goal was to construct a dedicated parallel computer system for molecular simulations, based on a ring architecture (since superseded by modern hardware designs). The molecular dynamics specific routines were rewritten in the programming language C from the Fortran 77-based program GROMOS, which had been developed in the same group.[ citation needed ]

Since 2001, GROMACS is developed by the GROMACS development teams at the Royal Institute of Technology and Uppsala University, Sweden.

Features

GROMACS is operated via the command-line interface, and can use files for input and output. It provides calculation progress and estimated time of arrival (ETA) feedback, a trajectory viewer, and an extensive library for trajectory analysis. [3] In addition, support for different force fields makes GROMACS very flexible. It can be executed in parallel, using Message Passing Interface (MPI) or threads. It contains a script to convert molecular coordinates from Protein Data Bank (PDB) files into the formats it uses internally. Once a configuration file for the simulation of several molecules (possibly including solvent) has been created, the simulation run (which can be time-consuming) produces a trajectory file, describing the movements of the atoms over time. That file can then be analyzed or visualized with several supplied tools. [10]

OpenCL and CUDA are possible for actual GPUs of AMD, Intel, and Nvidia with great acceleration against CPU based runs since Version 5 or higher. In Version 2021 OpenCL is deprecated and SYCL is in early new support. [11]

Easter eggs

As of January 2010, GROMACS' source code contains approximately 400 alternative backronyms to GROMACS as jokes among the developers and biochemistry researchers. These include "Gromacs Runs On Most of All Computer Systems", "Gromacs Runs One Microsecond At Cannonball Speeds", "Good ROcking Metal Altar for Chronical Sinner", "Working on GRowing Old MAkes el Chrono Sweat", and "Great Red Owns Many ACres of Sand". They are randomly selected to possibly appear in GROMACS's output stream. In one instance, such an bacronym, "Giving Russians Opium May Alter Current Situation", caused offense. [12]

Applications

Under a non-GPL license, GROMACS is widely used in the Folding@home distributed computing project for simulations of protein folding, where it is the base code for the project's largest and most regularly used series of calculation cores. [13] [14] EvoGrid, a distributed computing project to evolve artificial life, also employs GROMACS. [15]

See also

Related Research Articles

<span class="mw-page-title-main">AMBER</span>

Assisted Model Building with Energy Refinement (AMBER) is the name of a widely-used molecular dynamics software package originally developed by Peter Kollman's group at the University of California, San Francisco. It has also, subsequently, come to designate a family of force fields for molecular dynamics of biomolecules that can be used both within the AMBER software suite and with many modern computational platforms.

GROningen MOlecular Simulation (GROMOS) is the name of a force field for molecular dynamics simulation, and a related computer software package. Both are developed at the University of Groningen, and at the Computer-Aided Chemistry Group at the Laboratory for Physical Chemistry at the Swiss Federal Institute of Technology (ETH Zurich). At Groningen, Herman Berendsen was involved in its development.

<span class="mw-page-title-main">Folding@home</span> Distributed computing project simulating protein folding

Folding@home is a distributed computing project aimed to help scientists develop new therapeutics for a variety of diseases by the means of simulating protein dynamics. This includes the process of protein folding and the movements of proteins, and is reliant on simulations run on volunteers' personal computers. Folding@home is currently based at the University of Pennsylvania and led by Greg Bowman, a former student of Vijay Pande.

<span class="mw-page-title-main">Molecular mechanics</span> Use of classical mechanics to model molecular systems

Molecular mechanics uses classical mechanics to model molecular systems. The Born–Oppenheimer approximation is assumed valid and the potential energy of all systems is calculated as a function of the nuclear coordinates using force fields. Molecular mechanics can be used to study molecule systems ranging in size and complexity from small to large biological systems or material assemblies with many thousands to millions of atoms.

<span class="mw-page-title-main">Molecular modelling</span> Discovering chemical properties by physical simulations

Molecular modelling encompasses all methods, theoretical and computational, used to model or mimic the behaviour of molecules. The methods are used in the fields of computational chemistry, drug design, computational biology and materials science to study molecular systems ranging from small chemical systems to large biological molecules and material assemblies. The simplest calculations can be performed by hand, but inevitably computers are required to perform molecular modelling of any reasonably sized system. The common feature of molecular modelling methods is the atomistic level description of the molecular systems. This may include treating atoms as the smallest individual unit, or explicitly modelling protons and neutrons with its quarks, anti-quarks and gluons and electrons with its photons.

A chemical file format is a type of data file which is used specifically for depicting molecular data. One of the most widely used is the chemical table file format, which is similar to Structure Data Format (SDF) files. They are text files that represent multiple chemical structure records and associated data fields. The XYZ file format is a simple format that usually gives the number of atoms in the first line, a comment on the second, followed by a number of lines with atomic symbols and cartesian coordinates. The Protein Data Bank Format is commonly used for proteins but is also used for other types of molecules. There are many other types which are detailed below. Various software systems are available to convert from one format to another.

This is a list of computer programs that are predominantly used for molecular mechanics calculations.

<span class="mw-page-title-main">Molecular modeling on GPUs</span> Using graphics processing units for molecular simulations

Molecular modeling on GPU is the technique of using a graphics processing unit (GPU) for molecular simulations.

<span class="mw-page-title-main">MDynaMix</span>

Molecular Dynamics of Mixtures (MDynaMix) is a computer software package for general purpose molecular dynamics to simulate mixtures of molecules, interacting by AMBER- and CHARMM-like force fields in periodic boundary conditions. Algorithms are included for NVE, NVT, NPT, anisotropic NPT ensembles, and Ewald summation to treat electrostatic interactions. The code was written in a mix of Fortran 77 and 90. The package runs on Unix and Unix-like (Linux) workstations, clusters of workstations, and on Windows in sequential mode.

This is a list of notable computer programs that are used for nucleic acids simulations.

<span class="mw-page-title-main">Ascalaph Designer</span>

Ascalaph Designer is a computer program for general purpose molecular modelling for molecular design and simulations. It provides a graphical environment for the common programs of quantum and classical molecular modelling ORCA, NWChem, Firefly, CP2K and MDynaMix . The molecular mechanics calculations cover model building, energy optimizations and molecular dynamics. Firefly covers a wide range of quantum chemistry methods. Ascalaph Designer is free and open-source software, released under the GNU General Public License, version 2 (GPLv2).

OpenMM is a library for performing molecular dynamics simulations on a wide variety of hardware architectures. First released in January 2010, it was written by Peter Eastman at the Vijay S. Pande lab at Stanford University. It is notable for its implementation in the Folding@home project's core22 kernel. Core22, also developed at the Pande lab, uses OpenMM to perform protein dynamics simulations on GPUs via CUDA and OpenCL. During the COVID-19 pandemic, a peak of 280,000 GPUs were estimated to be running OpenMM via core22.

Local elevation is a technique used in computational chemistry or physics, mainly in the field of molecular simulation. It was developed in 1994 by Huber, Torda and van Gunsteren to enhance the searching of conformational space in molecular dynamics simulations and is available in the GROMOS software for molecular dynamics simulation. The method was, together with the conformational flooding method, the first to introduce memory dependence into molecular simulations. Many recent methods build on the principles of the local elevation technique, including the Engkvist-Karlström, adaptive biasing force, Wang–Landau, metadynamics, adaptively biased molecular dynamics, adaptive reaction coordinate forces, and local elevation umbrella sampling methods. The basic principle of the method is to add a memory-dependent potential energy term in the simulation so as to prevent the simulation to revisit already sampled configurations, which leads to the increased probability of discovering new configurations. The method can be seen as a continuous variant of the Tabu search method.

<span class="mw-page-title-main">Simon Fisher</span> British geneticist and neuroscientist (born 1970)

Simon E. Fisher is a British geneticist and neuroscientist who has pioneered research into the genetic basis of human speech and language. He is a director of the Max Planck Institute for Psycholinguistics and Professor of language and genetics at the Donders Institute for Brain, Cognition and Behaviour in Nijmegen, The Netherlands.

<span class="mw-page-title-main">SYCL</span> Higher-level programming standard for heterogeneous computing

SYCL is a higher-level programming model to improve programming productivity on various hardware accelerators. It is a single-source embedded domain-specific language (eDSL) based on pure C++17. It is a standard developed by Khronos Group, announced in March 2014.

MBN Explorer is a software package for molecular dynamics simulations, structure optimization and kinetic Monte Carlo simulations. It is designed for multiscale computational analysis of structure and dynamics of atomic clusters and nanoparticles, biomolecules and nanosystems, nanostructured materials, different states of matter and various interfaces. The software has been developed by MBN Research Center.

In the context of chemistry and molecular modelling, the Interface force field (IFF) is a force field for classical molecular simulations of atoms, molecules, and assemblies up to the large nanometer scale, covering compounds from across the periodic table. It employs a consistent classical Hamiltonian energy function for metals, oxides, and organic compounds, linking biomolecular and materials simulation platforms into a single platform. The reliability is often higher than that of density functional theory calculations at more than a million times lower computational cost. IFF includes a physical-chemical interpretation for all parameters as well as a surface model database that covers different cleavage planes and surface chemistry of included compounds. The Interface Force Field is compatible with force fields for the simulation of primarily organic compounds and can be used with common molecular dynamics and Monte Carlo codes. Structures and energies of included chemical elements and compounds are rigorously validated and property predictions are up to a factor of 100 more accurate relative to earlier models.

References

  1. "The GROMACS development team". Archived from the original on 2020-02-26. Retrieved 2012-06-27.
  2. "Gromacs Downloads". gromacs.org. Retrieved 2023-11-01.
  3. 1 2 3 "About Gromacs". gromacs.org. 16 August 2010. Archived from the original on 2020-11-27. Retrieved 2012-06-26.
  4. "People — Gromacs". gromacs.org. 14 March 2012. Archived from the original on 26 February 2020. Retrieved 26 June 2012.
  5. Van Der Spoel D, Lindahl E, Hess B, Groenhof G, Mark AE, Berendsen HJ (2005). "GROMACS: fast, flexible, and free". J Comput Chem. 26 (16): 1701–18. doi:10.1002/jcc.20291. PMID   16211538. S2CID   1231998.
  6. Hess B, Kutzner C, Van Der Spoel D, Lindahl E (2008). "GROMACS 4: Algorithms for Highly Efficient, Load-Balanced, and Scalable Molecular Simulation". J Chem Theory Comput. 4 (2): 435–447. doi:10.1021/ct700301q. hdl: 11858/00-001M-0000-0012-DDBF-0 . PMID   26620784. S2CID   1142192.
  7. Carsten Kutzner; David Van Der Spoel; Martin Fechner; Erik Lindahl; Udo W. Schmitt; Bert L. De Groot; Helmut Grubmüller (2007). "Speeding up parallel GROMACS on high-latency networks". Journal of Computational Chemistry. 28 (12): 2075–2084. doi:10.1002/jcc.20703. hdl: 11858/00-001M-0000-0012-E29A-0 . PMID   17405124. S2CID   519769.
  8. Berk Hess; Carsten Kutzner; David van der Spoel; Erik Lindahl (2008). "GROMACS 4: Algorithms for Highly Efficient, Load-Balanced, and Scalable Molecular Simulation". Journal of Chemical Theory and Computation. 4 (3): 435–447. doi:10.1021/ct700301q. hdl: 11858/00-001M-0000-0012-DDBF-0 . PMID   26620784. S2CID   1142192.
  9. "GPUs — Gromacs". gromacs.org. 20 January 2012. Archived from the original on 27 November 2020. Retrieved 26 June 2012.
  10. "GROMACS flow chart". gromacs.org. 18 January 2009. Archived from the original on 24 June 2010. Retrieved 26 June 2012.
  11. https://www.iwocl.org/wp-content/uploads/22-iwocl-syclcon-2021-alekseenko-slides.pdf [ bare URL PDF ]
  12. "Re: Working on Giving Russians Opium May Alter Current Situation". Folding@home. 17 January 2010. Retrieved 2012-06-26.
  13. Pande lab (11 June 2012). "Folding@home Open Source FAQ". Folding@home. Stanford University. Archived from the original (FAQ) on 17 July 2012. Retrieved 26 June 2012.
  14. Adam Beberg; Daniel Ensign; Guha Jayachandran; Siraj Khaliq; Vijay Pande (2009). "Folding@home: Lessons from eight years of volunteer distributed computing". 2009 IEEE International Symposium on Parallel & Distributed Processing (PDF). pp. 1–8. doi:10.1109/IPDPS.2009.5160922. ISBN   978-1-4244-3751-1. ISSN   1530-2075. S2CID   15677970.
  15. Markoff, John (29 September 2009). "Wanted: Home Computers to Join in Research on Artificial Life". The New York Times. Retrieved 26 June 2012.