The biochemical systems equation is a compact equation of nonlinear differential equations for describing a kinetic model for any network of coupled biochemical reactions and transport processes. [1] [2]
The equation is expressed in the following form:
The notation for the dependent variable x varies among authors. For example, some authors use s, indicating species. [2] x is used here to match the state space notation used in control theory but either notation is acceptable.
is the stoichiometry matrix which is an by matrix of stoichiometry coefficient. is the number of species and the number of biochemical reactions. The notation for is also variable. In constraint-based modeling the symbol tends to be used to indicate 'stoichiometry'. However in biochemical dynamic modeling [3] and sensitivity analysis, tends to be in more common use to indicate 'number'. In the chemistry domain, the symbol used for the stoichiometry matrix is highly variable though the symbols S and N have been used in the past. [4] [5]
is an n-dimensional column vector of reaction rates, and is a p-dimensional column vector of parameters.
Given the biochemical network:
where and are fixed species to ensure the system is open. The system equation can be written as: [1] [6]
So that:
The elements of the rate vector will be rate equations that are functions of one or more species and parameters, p. In the example, these might be simple mass-action rate laws such as where is the rate constant parameter. The particular laws chosen will depend on the specific system under study. Assuming mass-action kinetics, the above equation can be written in complete form as:
The system equation can be analyzed by looking at the linear response of the equation around the steady-state with respect to the parameter . [7] At steady-state, the system equation is set to zero and given by:
Differentiating the equation with respect to and rearranging gives:
This derivation assumes that the stoichiometry matrix has full rank. If this is not the case, then the inverse won't exist.
For example, consider the same problem from the previous section of a linear chain. The matrix is the unscaled elasticity matrix:
In this specific problem there are 3 species () and 4 reaction steps (), the elasticity matrix is therefore a matrix. However, a number of entries in the matrix will be zero. For example will be zero since has no effect on . The matrix, therefore, will contain the following entries:
The parameter matrix depends on which parameters are considered. In Metabolic control analysis, a common set of parameters are the enzyme activities. For the sake of argument, we can equate the rate constants with the enzyme activity parameters. We also assume that each enzyme, , only can affect its own step and no other. The matrix is the unscaled elasticity matrix with respect to the parameters. Since there are 4 reaction steps and 4 corresponding parameters, the matrix will be a 4 by 4 matrix. Since each parameter only affects one reaction, the matrix will be a diagonal matrix:
Since there are 3 species and 4 reactions, the resulting matrix will be a 3 by 4 matrix
Each expression in the matrix describes how a given parameter influences the steady-state concentration of a given species. Note that this is the unscaled derivative. It is often the case that the derivative is scaled by the parameter and concentration to eliminate units as well as turn the measure into a relative change.
The biochemical systems equation makes two key assumptions:
In vector calculus, divergence is a vector operator that operates on a vector field, producing a scalar field giving the quantity of the vector field's source at each point. More technically, the divergence represents the volume density of the outward flux of a vector field from an infinitesimal volume around a given point.
In physics, the Lorentz transformations are a six-parameter family of linear transformations from a coordinate frame in spacetime to another frame that moves at a constant velocity relative to the former. The respective inverse transformation is then parameterized by the negative of this velocity. The transformations are named after the Dutch physicist Hendrik Lorentz.
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a function whose value at any given sample in the sample space can be interpreted as providing a relative likelihood that the value of the random variable would be equal to that sample. Probability density is the probability per unit length, in other words, while the absolute likelihood for a continuous random variable to take on any particular value is 0, the value of the PDF at two different samples can be used to infer, in any particular draw of the random variable, how much more likely it is that the random variable would be close to one sample compared to the other sample.
In physics, angular velocity, also known as angular frequency vector, is a pseudovector representation of how fast the angular position or orientation of an object changes with time. The magnitude of the pseudovector represents the angular speed, the rate at which the object rotates or revolves, and its direction is normal to the instantaneous plane of rotation or angular displacement. The orientation of angular velocity is conventionally specified by the right-hand rule.
In physics, a Langevin equation is a stochastic differential equation describing how a system evolves when subjected to a combination of deterministic and fluctuating ("random") forces. The dependent variables in a Langevin equation typically are collective (macroscopic) variables changing only slowly in comparison to the other (microscopic) variables of the system. The fast (microscopic) variables are responsible for the stochastic nature of the Langevin equation. One application is to Brownian motion, which models the fluctuating motion of a small particle in a fluid.
In statistics, the Gauss–Markov theorem states that the ordinary least squares (OLS) estimator has the lowest sampling variance within the class of linear unbiased estimators, if the errors in the linear regression model are uncorrelated, have equal variances and expectation value of zero. The errors do not need to be normal, nor do they need to be independent and identically distributed. The requirement that the estimator be unbiased cannot be dropped, since biased estimators exist with lower variance. See, for example, the James–Stein estimator, ridge regression, or simply any degenerate estimator.
In continuum mechanics, the infinitesimal strain theory is a mathematical approach to the description of the deformation of a solid body in which the displacements of the material particles are assumed to be much smaller than any relevant dimension of the body; so that its geometry and the constitutive properties of the material at each point of space can be assumed to be unchanged by the deformation.
In vector calculus, the Jacobian matrix of a vector-valued function of several variables is the matrix of all its first-order partial derivatives. When this matrix is square, that is, when the function takes the same number of variables as input as the number of vector components of its output, its determinant is referred to as the Jacobian determinant. Both the matrix and the determinant are often referred to simply as the Jacobian in literature.
In the mathematical field of differential geometry, a metric tensor is an additional structure on a manifold M that allows defining distances and angles, just as the inner product on a Euclidean space allows defining distances and angles there. More precisely, a metric tensor at a point p of M is a bilinear form defined on the tangent space at p, and a metric tensor on M consists of a metric tensor at each point p of M that varies smoothly with p.
In mathematics, the Hessian matrix, Hessian or Hesse matrix is a square matrix of second-order partial derivatives of a scalar-valued function, or scalar field. It describes the local curvature of a function of many variables. The Hessian matrix was developed in the 19th century by the German mathematician Ludwig Otto Hesse and later named after him. Hesse originally used the term "functional determinants".
In mathematical statistics, the Fisher information is a way of measuring the amount of information that an observable random variable X carries about an unknown parameter θ of a distribution that models X. Formally, it is the variance of the score, or the expected value of the observed information.
In geometry, curvilinear coordinates are a coordinate system for Euclidean space in which the coordinate lines may be curved. These coordinates may be derived from a set of Cartesian coordinates by using a transformation that is locally invertible at each point. This means that one can convert a point given in a Cartesian coordinate system to its curvilinear coordinates and back. The name curvilinear coordinates, coined by the French mathematician Lamé, derives from the fact that the coordinate surfaces of the curvilinear systems are curved.
In mathematics, matrix calculus is a specialized notation for doing multivariable calculus, especially over spaces of matrices. It collects the various partial derivatives of a single function with respect to many variables, and/or of a multivariate function with respect to a single variable, into vectors and matrices that can be treated as single entities. This greatly simplifies operations such as finding the maximum or minimum of a multivariate function and solving systems of differential equations. The notation used here is commonly used in statistics and engineering, while the tensor index notation is preferred in physics.
In electromagnetism, the electromagnetic tensor or electromagnetic field tensor is a mathematical object that describes the electromagnetic field in spacetime. The field tensor was first used after the four-dimensional tensor formulation of special relativity was introduced by Hermann Minkowski. The tensor allows related physical laws to be written very concisely, and allows for the quantization of the electromagnetic field by Lagrangian formulation described below.
In applied mathematics, in particular the context of nonlinear system analysis, a phase plane is a visual display of certain characteristics of certain kinds of differential equations; a coordinate plane with axes being the values of the two state variables, say (x, y), or (q, p) etc. (any pair of variables). It is a two-dimensional case of the general n-dimensional phase space.
The Cauchy momentum equation is a vector partial differential equation put forth by Cauchy that describes the non-relativistic momentum transport in any continuum.
In numerical analysis, the interval finite element method is a finite element method that uses interval parameters. Interval FEM can be applied in situations where it is not possible to get reliable probabilistic characteristics of the structure. This is important in concrete structures, wood structures, geomechanics, composite structures, biomechanics and in many other areas. The goal of the Interval Finite Element is to find upper and lower bounds of different characteristics of the model and use these results in the design process. This is so called worst case design, which is closely related to the limit state design.
A differential equation is a mathematical equation for an unknown function of one or several variables that relates the values of the function itself and its derivatives of various orders. A matrix differential equation contains more than one function stacked into vector form with a matrix relating the functions to their derivatives.
In mathematical biology, the community matrix is the linearization of a generalized Lotka–Volterra equation at an equilibrium point. The eigenvalues of the community matrix determine the stability of the equilibrium point.
Lie point symmetry is a concept in advanced mathematics. Towards the end of the nineteenth century, Sophus Lie introduced the notion of Lie group in order to study the solutions of ordinary differential equations (ODEs). He showed the following main property: the order of an ordinary differential equation can be reduced by one if it is invariant under one-parameter Lie group of point transformations. This observation unified and extended the available integration techniques. Lie devoted the remainder of his mathematical career to developing these continuous groups that have now an impact on many areas of mathematically based sciences. The applications of Lie groups to differential systems were mainly established by Lie and Emmy Noether, and then advocated by Élie Cartan.