Differential-algebraic system of equations

Last updated

In mathematics, a differential-algebraic system of equations (DAE) is a system of equations that either contains differential equations and algebraic equations, or is equivalent to such a system.

Contents

The set of the solutions of such a system is a differential algebraic variety, and corresponds to an ideal in a differential algebra of differential polynomials.

In the univariate case, a DAE in the variable t can be written as a single equation of the form

where is a vector of unknown functions and the overdot denotes the time derivative, i.e., .

They are distinct from ordinary differential equation (ODE) in that a DAE is not completely solvable for the derivatives of all components of the function x because these may not all appear (i.e. some equations are algebraic); technically the distinction between an implicit ODE system [that may be rendered explicit] and a DAE system is that the Jacobian matrix is a singular matrix for a DAE system. [1] This distinction between ODEs and DAEs is made because DAEs have different characteristics and are generally more difficult to solve. [2]

In practical terms, the distinction between DAEs and ODEs is often that the solution of a DAE system depends on the derivatives of the input signal and not just the signal itself as in the case of ODEs; [3] this issue is commonly encountered in nonlinear systems with hysteresis, [4] such as the Schmitt trigger. [5]

This difference is more clearly visible if the system may be rewritten so that instead of x we consider a pair of vectors of dependent variables and the DAE has the form

where , , and

A DAE system of this form is called semi-explicit. [1] Every solution of the second half g of the equation defines a unique direction for x via the first half f of the equations, while the direction for y is arbitrary. But not every point (x,y,t) is a solution of g. The variables in x and the first half f of the equations get the attribute differential. The components of y and the second half g of the equations are called the algebraic variables or equations of the system. [The term algebraic in the context of DAEs only means free of derivatives and is not related to (abstract) algebra.]

The solution of a DAE consists of two parts, first the search for consistent initial values and second the computation of a trajectory. To find consistent initial values it is often necessary to consider the derivatives of some of the component functions of the DAE. The highest order of a derivative that is necessary for this process is called the differentiation index. The equations derived in computing the index and consistent initial values may also be of use in the computation of the trajectory. A semi-explicit DAE system can be converted to an implicit one by decreasing the differentiation index by one, and vice versa. [6]

Other forms of DAEs

The distinction of DAEs to ODEs becomes apparent if some of the dependent variables occur without their derivatives. The vector of dependent variables may then be written as pair and the system of differential equations of the DAE appears in the form

where

As a whole, the set of DAEs is a function

Initial conditions must be a solution of the system of equations of the form

Examples

The behaviour of a pendulum of length L with center in (0,0) in Cartesian coordinates (x,y) is described by the Euler–Lagrange equations

where is a Lagrange multiplier. The momentum variables u and v should be constrained by the law of conservation of energy and their direction should point along the circle. Neither condition is explicit in those equations. Differentiation of the last equation leads to

restricting the direction of motion to the tangent of the circle. The next derivative of this equation implies

and the derivative of that last identity simplifies to which implies the conservation of energy since after integration the constant is the sum of kinetic and potential energy.

To obtain unique derivative values for all dependent variables the last equation was three times differentiated. This gives a differentiation index of 3, which is typical for constrained mechanical systems.

If initial values and a sign for y are given, the other variables are determined via , and if then and . To proceed to the next point it is sufficient to get the derivatives of x and u, that is, the system to solve is now

This is a semi-explicit DAE of index 1. Another set of similar equations may be obtained starting from and a sign for x.

DAEs also naturally occur in the modelling of circuits with non-linear devices. Modified nodal analysis employing DAEs is used for example in the ubiquitous SPICE family of numeric circuit simulators. [7] Similarly, Fraunhofer's Analog Insydes Mathematica package can be used to derive DAEs from a netlist and then simplify or even solve the equations symbolically in some cases. [8] [9] It is worth noting that the index of a DAE (of a circuit) can be made arbitrarily high by cascading/coupling via capacitors operational amplifiers with positive feedback. [4]

Semi-explicit DAE of index 1

DAE of the form

are called semi-explicit. The index-1 property requires that g is solvable for y. In other words, the differentiation index is 1 if by differentiation of the algebraic equations for t an implicit ODE system results,

which is solvable for if

Every sufficiently smooth DAE is almost everywhere reducible to this semi-explicit index-1 form.

Numerical treatment of DAE and applications

Two major problems in solving DAEs are index reduction and consistent initial conditions. Most numerical solvers require ordinary differential equations and algebraic equations of the form

It is a non-trivial task to convert arbitrary DAE systems into ODEs for solution by pure ODE solvers. Techniques which can be employed include Pantelides algorithm and dummy derivative index reduction method . Alternatively, a direct solution of high-index DAEs with inconsistent initial conditions is also possible. This solution approach involves a transformation of the derivative elements through orthogonal collocation on finite elements or direct transcription into algebraic expressions. This allows DAEs of any index to be solved without rearrangement in the open equation form

Once the model has been converted to algebraic equation form, it is solvable by large-scale nonlinear programming solvers (see APMonitor).

Tractability

Several measures of DAEs tractability in terms of numerical methods have developed, such as differentiation index, perturbation index, tractability index, geometric index, and the Kronecker index. [10] [11]

Structural analysis for DAEs

We use the -method to analyze a DAE. We construct for the DAE a signature matrix , where each row corresponds to each equation and each column corresponds to each variable . The entry in position is , which denotes the highest order of derivative to which occurs in , or if does not occur in .

For the pendulum DAE above, the variables are . The corresponding signature matrix is

See also

Related Research Articles

In particle physics, the Dirac equation is a relativistic wave equation derived by British physicist Paul Dirac in 1928. In its free form, or including electromagnetic interactions, it describes all spin-1/2 massive particles, called "Dirac particles", such as electrons and quarks for which parity is a symmetry. It is consistent with both the principles of quantum mechanics and the theory of special relativity, and was the first theory to account fully for special relativity in the context of quantum mechanics. It was validated by accounting for the fine structure of the hydrogen spectrum in a completely rigorous way. It has become vital in the building of the Standard Model.

In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function subject to equation constraints. It is named after the mathematician Joseph-Louis Lagrange.

<span class="mw-page-title-main">Heat equation</span> Partial differential equation describing the evolution of temperature in a region

In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for the purpose of modeling how a quantity such as heat diffuses through a given region.

<span class="mw-page-title-main">Optimal control</span> Mathematical way of attaining a desired output from a dynamic system

Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations research. For example, the dynamical system might be a spacecraft with controls corresponding to rocket thrusters, and the objective might be to reach the Moon with minimum fuel expenditure. Or the dynamical system could be a nation's economy, with the objective to minimize unemployment; the controls in this case could be fiscal and monetary policy. A dynamical system may also be introduced to embed operations research problems within the framework of optimal control theory.

<span class="mw-page-title-main">Separation of variables</span> Technique for solving differential equations

In mathematics, separation of variables is any of several methods for solving ordinary and partial differential equations, in which algebra allows one to rewrite an equation so that each of two variables occurs on a different side of the equation.

In mathematics, integral equations are equations in which an unknown function appears under an integral sign. In mathematical notation, integral equations may thus be expressed as being of the form:

In mathematics, the Schwarzian derivative is an operator similar to the derivative which is invariant under Möbius transformations. Thus, it occurs in the theory of the complex projective line, and in particular, in the theory of modular forms and hypergeometric functions. It plays an important role in the theory of univalent functions, conformal mapping and Teichmüller spaces. It is named after the German mathematician Hermann Schwarz.

In mathematics, the method of characteristics is a technique for solving partial differential equations. Typically, it applies to first-order equations, although more generally the method of characteristics is valid for any hyperbolic and parabolic partial differential equation. The method is to reduce a partial differential equation to a family of ordinary differential equations along which the solution can be integrated from some initial data given on a suitable hypersurface.

In mathematics, an Euler–Cauchy equation, or Cauchy–Euler equation, or simply Euler's equation is a linear homogeneous ordinary differential equation with variable coefficients. It is sometimes referred to as an equidimensional equation. Because of its particularly simple equidimensional structure, the differential equation can be solved explicitly.

In linear algebra, it is often important to know which vectors have their directions unchanged by a given linear transformation. An eigenvector or characteristic vector is such a vector. More precisely, an eigenvector of a linear transformation is scaled by a constant factor when the linear transformation is applied to it: . The corresponding eigenvalue, characteristic value, or characteristic root is the multiplying factor .

In mathematics, delay differential equations (DDEs) are a type of differential equation in which the derivative of the unknown function at a certain time is given in terms of the values of the function at previous times. DDEs are also called time-delay systems, systems with aftereffect or dead-time, hereditary systems, equations with deviating argument, or differential-difference equations. They belong to the class of systems with the functional state, i.e. partial differential equations (PDEs) which are infinite dimensional, as opposed to ordinary differential equations (ODEs) having a finite dimensional state vector. Four points may give a possible explanation of the popularity of DDEs:

  1. Aftereffect is an applied problem: it is well known that, together with the increasing expectations of dynamic performances, engineers need their models to behave more like the real process. Many processes include aftereffect phenomena in their inner dynamics. In addition, actuators, sensors, and communication networks that are now involved in feedback control loops introduce such delays. Finally, besides actual delays, time lags are frequently used to simplify very high order models. Then, the interest for DDEs keeps on growing in all scientific areas and, especially, in control engineering.
  2. Delay systems are still resistant to many classical controllers: one could think that the simplest approach would consist in replacing them by some finite-dimensional approximations. Unfortunately, ignoring effects which are adequately represented by DDEs is not a general alternative: in the best situation, it leads to the same degree of complexity in the control design. In worst cases, it is potentially disastrous in terms of stability and oscillations.
  3. Voluntary introduction of delays can benefit the control system.
  4. In spite of their complexity, DDEs often appear as simple infinite-dimensional models in the very complex area of partial differential equations (PDEs).

A differential equation can be homogeneous in either of two respects.

<span class="mw-page-title-main">Elliptic boundary value problem</span>

In mathematics, an elliptic boundary value problem is a special kind of boundary value problem which can be thought of as the stable state of an evolution problem. For example, the Dirichlet problem for the Laplacian gives the eventual distribution of heat in a room several hours after the heating is turned on.

A differential equation is a mathematical equation for an unknown function of one or several variables that relates the values of the function itself and its derivatives of various orders. A matrix differential equation contains more than one function stacked into vector form with a matrix relating the functions to their derivatives.

<span class="mw-page-title-main">Lie point symmetry</span>

Lie point symmetry is a concept in advanced mathematics. Towards the end of the nineteenth century, Sophus Lie introduced the notion of Lie group in order to study the solutions of ordinary differential equations (ODEs). He showed the following main property: the order of an ordinary differential equation can be reduced by one if it is invariant under one-parameter Lie group of point transformations. This observation unified and extended the available integration techniques. Lie devoted the remainder of his mathematical career to developing these continuous groups that have now an impact on many areas of mathematically based sciences. The applications of Lie groups to differential systems were mainly established by Lie and Emmy Noether, and then advocated by Élie Cartan.

<span class="mw-page-title-main">Ordinary differential equation</span> Differential equation containing derivatives with respect to only one variable

In mathematics, an ordinary differential equation (ODE) is a differential equation (DE) dependent on only a single independent variable. As with other DE, its unknown(s) consists of one function(s) and involves the derivatives of those functions. The term "ordinary" is used in contrast with partial differential equations (PDEs) which may be with respect to more than one independent variable, and, less commonly, in contrast with stochastic differential equations (SDEs) where the progression is random.

In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.

The Fokas method, or unified transform, is an algorithmic procedure for analysing boundary value problems for linear partial differential equations and for an important class of nonlinear PDEs belonging to the so-called integrable systems. It is named after Greek mathematician Athanassios S. Fokas.

The Segregated Runge–Kutta (SRK) method is a family of IMplicit–EXplicit (IMEX) Runge–Kutta methods that were developed to approximate the solution of differential algebraic equations (DAE) of index 2.

Tau functions are an important ingredient in the modern mathematical theory of integrable systems, and have numerous applications in a variety of other domains. They were originally introduced by Ryogo Hirota in his direct method approach to soliton equations, based on expressing them in an equivalent bilinear form.

References

  1. 1 2 Uri M. Ascher; Linda R. Petzold (1998). Computer Methods for Ordinary Differential Equations and Differential-Algebraic Equations. SIAM. p. 12. ISBN   978-1-61197-139-2.
  2. Achim Ilchmann; Timo Reis (2014). Surveys in Differential-Algebraic Equations II. Springer. pp. 104–105. ISBN   978-3-319-11050-9.
  3. Renate Merker; Wolfgang Schwarz, eds. (2001). System Design Automation: Fundamentals, Principles, Methods, Examples . Springer Science & Business Media. p.  221. ISBN   978-0-7923-7313-1.
  4. 1 2 K. E. Brenan; S. L. Campbell; L. R. Petzold (1996). Numerical Solution of Initial-value Problems in Differential-algebraic Equations. SIAM. pp. 173–177. ISBN   978-1-61197-122-4.
  5. Günther, M.; Feldmann, U.; Ter Maten, J. (2005). "Modelling and Discretization of Circuit Problems". Numerical Methods in Electromagnetics. Handbook of Numerical Analysis. Vol. 13. p. 523. doi:10.1016/S1570-8659(04)13006-8. ISBN   978-0-444-51375-5., pp. 529-531
  6. Ascher and Petzold, p. 234
  7. Ricardo Riaza (2013). "DAEs in Circuit Modelling: A Survey". In Achim Ilchmann; Timo Reis (eds.). Surveys in Differential-Algebraic Equations I. Springer Science & Business Media. ISBN   978-3-642-34928-7.
  8. Platte, D.; Jing, S.; Sommer, R.; Barke, E. (2007). "Improving Efficiency and Robustness of Analog Behavioral Models". Advances in Design and Specification Languages for Embedded Systems. p. 53. doi:10.1007/978-1-4020-6149-3_4. ISBN   978-1-4020-6147-9.
  9. Hauser, M.; Salzig, C.; Dreyer, A. (2011). "Fast and Robust Symbolic Model Order Reduction with Analog Insydes". Computer Algebra in Scientific Computing. Lecture Notes in Computer Science. Vol. 6885. p. 215. doi:10.1007/978-3-642-23568-9_17. ISBN   978-3-642-23567-2.
  10. Ricardo Riaza (2008). Differential-algebraic Systems: Analytical Aspects and Circuit Applications . World Scientific. pp.  5–8. ISBN   978-981-279-181-8.
  11. Takamatsu, Mizuyo; Iwata, Satoru (2008). "Index characterization of differential-algebraic equations in hybrid analysis for circuit simulation" (PDF). International Journal of Circuit Theory and Applications . 38 (4): 419–440. doi:10.1002/cta.577. S2CID   3875504. Archived from the original (PDF) on 16 December 2014. Retrieved 9 November 2022.

Further reading

Books

Various papers