Pfaffian function

Last updated

In mathematics, Pfaffian functions are a certain class of functions whose derivative can be written in terms of the original function. They were originally introduced by Askold Khovanskii in the 1970s, but are named after German mathematician Johann Pfaff.

Contents

Basic definition

Some functions, when differentiated, give a result which can be written in terms of the original function. Perhaps the simplest example is the exponential function, f(x) = ex. If we differentiate this function we get ex again, that is

Another example of a function like this is the reciprocal function, g(x) = 1/x. If we differentiate this function we will see that

Other functions may not have the above property, but their derivative may be written in terms of functions like those above. For example, if we take the function h(x) = ex log(x) then we see

Functions like these form the links in a so-called Pfaffian chain. Such a chain is a sequence of functions, say f1, f2, f3, etc., with the property that if we differentiate any of the functions in this chain then the result can be written in terms of the function itself and all the functions preceding it in the chain (specifically as a polynomial in those functions and the variables involved). So with the functions above we have that f, g, h is a Pfaffian chain.

A Pfaffian function is then just a polynomial in the functions appearing in a Pfaffian chain and the function argument. So with the Pfaffian chain just mentioned, functions such as F(x) = x3f(x)2  2g(x)h(x) are Pfaffian.

Rigorous definition

Let U be an open domain in Rn. A Pfaffian chain of order r  0 and degree α  1 in U is a sequence of real analytic functions f1,..., fr in U satisfying differential equations

for i = 1, ..., r where Pi, j  R[x1, ..., xn, y1, ..., yi] are polynomials of degreeα. A function f on U is called a Pfaffian function of order r and degree (α, β) if

where P  R[x1, ..., xn, y1, ..., yr] is a polynomial of degree at most β ≥ 1. The numbers r, α, and β are collectively known as the format of the Pfaffian function, and give a useful measure of its complexity.

Examples

In model theory

Consider the structure R = (R, +, −, ·, <, 0, 1), the ordered field of real numbers. In the 1960s Andrei Gabrielov proved that the structure obtained by starting with R and adding a function symbol for every analytic function restricted to the unit box [0, 1]m is model complete. [2] That is, any set definable in this structure Ran was just the projection of some higher-dimensional set defined by identities and inequalities involving these restricted analytic functions.

In the 1990s, Alex Wilkie showed that one has the same result if instead of adding every restricted analytic function, one just adds the unrestricted exponential function to R to get the ordered real field with exponentiation, Rexp, a result known as Wilkie's theorem. [3] Wilkie also tackled the question of which finite sets of analytic functions could be added to R to get a model-completeness result. It turned out that adding any Pfaffian chain restricted to the box [0, 1]m would give the same result. In particular one may add all Pfaffian functions to R to get the structure RPfaff as a variant of Gabrielov's result. The result on exponentiation is not a special case of this result (even though exp is a Pfaffian chain by itself), as it applies to the unrestricted exponential function.

This result of Wilkie's proved that the structure RPfaff is an o-minimal structure.

Noetherian functions

The equations above that define a Pfaffian chain are said to satisfy a triangular condition, since the derivative of each successive function in the chain is a polynomial in one extra variable. Thus if they are written out in turn a triangular shape appears:

and so on. If this triangularity condition is relaxed so that the derivative of each function in the chain is a polynomial in all the other functions in the chain, then the chain of functions is known as a Noetherian chain, and a function constructed as a polynomial in this chain is called a Noetherian function. [4] So, for example, a Noetherian chain of order three is composed of three functions f1, f2, f3, satisfying the equations

The name stems from the fact that the ring generated by the functions in such a chain is Noetherian. [5]

Any Pfaffian chain is also a Noetherian chain; the extra variables in each polynomial are simply redundant in this case. But not every Noetherian chain is Pfaffian; for example, if we take f1(x) = sin(x) and f2(x) = cos(x) then we have the equations

and these hold for all real numbers x, so f1, f2 is a Noetherian chain on all of R. But there is no polynomial P(x, y) such that the derivative of sin(x) can be written as P(x, sin(x)), and so this chain is not Pfaffian.

Notes

  1. Liouville functions are essentially all the real analytic functions obtainable from the elementary functions by applying the usual arithmetic operations, exponentiation, and integration. They are unrelated to Liouville's function in number theory.
  2. A. Gabrielov, "Projections of semi-analytic sets", Functional Anal. Appl.2 (1968), pp.282–291.
  3. A.J. Wilkie, "Model completeness results for expansions of the ordered field of real numbers by restricted Pfaffian functions and the exponential functions", J. Amer. Math. Soc.9 (1996), pp. 1051–1094.
  4. Andrei Gabrielov, Nicolai Vorobjov (2004). "Complexity of computations with Pfaffian and Noetherian functions". In Yulij Ilyashenko, Christiane Rousseau (ed.). Normal Forms, Bifurcations and Finiteness Problems in Differential Equations. Kluwer Academic Publishers. ISBN   1-4020-1928-9.
  5. J.C. Tougeron, "Algèbres analytiques topologiquement nœthériennes, Théorie de Hovanskii", Annales de l'Institut Fourier41 (1991), pp.823840.

Related Research Articles

<span class="mw-page-title-main">Derivative</span> Instantaneous rate of change (mathematics)

In mathematics, the derivative shows the sensitivity of change of a function's output with respect to the input. Derivatives are a fundamental tool of calculus. For example, the derivative of the position of a moving object with respect to time is the object's velocity: this measures how quickly the position of the object changes when time advances.

<span class="mw-page-title-main">Taylor's theorem</span> Approximation of a function by a truncated power series

In calculus, Taylor's theorem gives an approximation of a -times differentiable function around a given point by a polynomial of degree , called the -th-order Taylor polynomial. For a smooth function, the Taylor polynomial is the truncation at the order of the Taylor series of the function. The first-order Taylor polynomial is the linear approximation of the function, and the second-order Taylor polynomial is often referred to as the quadratic approximation. There are several versions of Taylor's theorem, some giving explicit estimates of the approximation error of the function by its Taylor polynomial.

<span class="mw-page-title-main">Least squares</span> Approximation method in statistics

The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems by minimizing the sum of the squares of the residuals made in the results of each individual equation.

In the calculus of variations and classical mechanics, the Euler–Lagrange equations are a system of second-order ordinary differential equations whose solutions are stationary points of the given action functional. The equations were discovered in the 1750s by Swiss mathematician Leonhard Euler and Italian mathematician Joseph-Louis Lagrange.

In mathematics, a transcendental function is an analytic function that does not satisfy a polynomial equation, in contrast to an algebraic function. In other words, a transcendental function "transcends" algebra in that it cannot be expressed algebraically using a finite amount of terms.

In the calculus of variations, a field of mathematical analysis, the functional derivative relates a change in a functional to a change in a function on which the functional depends.

<span class="mw-page-title-main">Quartic function</span> Polynomial function of degree four

In algebra, a quartic function is a function of the form

In mathematics, the classical orthogonal polynomials are the most widely used orthogonal polynomials: the Hermite polynomials, Laguerre polynomials, Jacobi polynomials.

A directional derivative is a concept in multivariable calculus that measures the rate at which a function changes in a particular direction at a given point.

<span class="mw-page-title-main">Gauss–Newton algorithm</span> Mathematical algorithm

The Gauss–Newton algorithm is used to solve non-linear least squares problems, which is equivalent to minimizing a sum of squared function values. It is an extension of Newton's method for finding a minimum of a non-linear function. Since a sum of squares must be nonnegative, the algorithm can be viewed as using Newton's method to iteratively approximate zeroes of the components of the sum, and thus minimizing the sum. In this sense, the algorithm is also an effective method for solving overdetermined systems of equations. It has the advantage that second derivatives, which can be challenging to compute, are not required.

<span class="mw-page-title-main">Hypergeometric function</span> Special function defined by a hypergeometric series

In mathematics, the Gaussian or ordinary hypergeometric function2F1(a,b;c;z) is a special function represented by the hypergeometric series, that includes many other special functions as specific or limiting cases. It is a solution of a second-order linear ordinary differential equation (ODE). Every second-order linear ODE with three regular singular points can be transformed into this equation.

<span class="mw-page-title-main">Differentiable manifold</span> Manifold upon which it is possible to perform calculus

In mathematics, a differentiable manifold is a type of manifold that is locally similar enough to a vector space to allow one to apply calculus. Any manifold can be described by a collection of charts (atlas). One may then apply ideas from calculus while working within the individual charts, since each chart lies within a vector space to which the usual rules of calculus apply. If the charts are suitably compatible, then computations done in one chart are valid in any other differentiable chart.

The Adomian decomposition method (ADM) is a semi-analytical method for solving ordinary and partial nonlinear differential equations. The method was developed from the 1970s to the 1990s by George Adomian, chair of the Center for Applied Mathematics at the University of Georgia. It is further extensible to stochastic systems by using the Ito integral. The aim of this method is towards a unified theory for the solution of partial differential equations (PDE); an aim which has been superseded by the more general theory of the homotopy analysis method. The crucial aspect of the method is employment of the "Adomian polynomials" which allow for solution convergence of the nonlinear portion of the equation, without simply linearizing the system. These polynomials mathematically generalize to a Maclaurin series about an arbitrary external parameter; which gives the solution method more flexibility than direct Taylor series expansion.

In classical mechanics, holonomic constraints are relations between the position variables that can be expressed in the following form:

Non-linear least squares is the form of least squares analysis used to fit a set of m observations with a model that is non-linear in n unknown parameters (m ≥ n). It is used in some forms of nonlinear regression. The basis of the method is to approximate the model by a linear one and to refine the parameters by successive iterations. There are many similarities to linear least squares, but also some significant differences. In economic theory, the non-linear least squares method is applied in (i) the probit regression, (ii) threshold regression, (iii) smooth regression, (iv) logistic link regression, (v) Box–Cox transformed regressors ().

In model theory, Tarski's exponential function problem asks whether the theory of the real numbers together with the exponential function is decidable. Alfred Tarski had previously shown that the theory of the real numbers is decidable.

In mathematics, Wilkie's theorem is a result by Alex Wilkie about the theory of ordered fields with an exponential function, or equivalently about the geometric nature of exponential varieties.

The purpose of this page is to provide supplementary materials for the ordinary least squares article, reducing the load of the main article with mathematics and improving its accessibility, while at the same time retaining the completeness of exposition.

Linear least squares (LLS) is the least squares approximation of linear functions to data. It is a set of formulations for solving statistical problems involved in linear regression, including variants for ordinary (unweighted), weighted, and generalized (correlated) residuals. Numerical methods for linear least squares include inverting the matrix of the normal equations and orthogonal decomposition methods.

In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.

References