Coopmans approximation

Last updated

The Coopmans approximation is a method for approximating a fractional-order integrator in a continuous process with constant space complexity. The most correct and accurate methods for calculating the fractional integral require a record of all previous history, and therefore would require a linear space complexity solution O(n), where n is the number of samples measured for the complete history.

A fractional-order integrator or just simply fractional integrator is an integrator device that calculates the fractional-order integral or derivative of an input. Differentiation or integration is a real or complex parameter. The fractional integrator is useful in fractional-order control where the history of the system under control is important to the control system output.

The fractor ( fractional capacitor ) is an analog component useful in control systems. In order to model the components behavior in a digital simulation, or replace the fractor in a digital controller, a linear solution is untenable. In order to reduce the space complexity however, it is necessary to lose information in some way.

Control system system to control other devices using control loops

A control system manages, commands, directs, or regulates the behavior of other devices or systems using control loops. It can range from a single home heating controller using a thermostat controlling a domestic boiler to large Industrial control systems which are used for controlling processes or machines.

The Coopmans approximation is a robust, simple method that uses a simple convolution to compute the fractional integral, then recycles old data back through the convolution. The convolution sets up a weighting table as described by the fractional calculus, which varies based on the size of the table, the sampling rate of the system, and the order of the integral. Once computed the weighting table remains static.

Convolution mathematical operation

In mathematics convolution is a mathematical operation on two functions to produce a third function that expresses how the shape of one is modified by the other. The term convolution refers to both the result function and to the process of computing it. Convolution is similar to cross-correlation. For real-valued functions, of a continuous or discrete variable, it differs from cross-correlation only in that either f (x) or g(x) is reflected about the y-axis; thus it is a cross-correlation of f (x) and g(−x), or f (−x) and g(x). For continuous functions, the cross-correlation operator is the adjoint of the convolution operator.

Fractional calculus is a branch of mathematical analysis that studies the several different possibilities of defining real number powers or complex number powers of the differentiation operator D

The data table is initialized as all zeros, which represents a lack of activity for all previous time. New data is added to the data buffer in the fashion of a ring buffer, so that the newest point is written over the oldest data point. The convolution is solved by multiplying corresponding elements from the weight and data tables, and summing the resulting products. As described, the loss of the old data by overwriting with new data will cause echoes in a continuous system as disturbances that were absorbed into the system are suddenly removed.

The solution to this is the crux of the Coopmans approximation, where the old data point, multiplied by its corresponding weight term, is added to the newest data point directly. This allows a smooth (though exponential, rather than power law) decay of the system history. This approximation has the desirable effect of removing the echo, while preserving the space complexity of the solution.

The negative effect of the approximation is that the phase character of the solution is lost as the system frequency approaches DC. However, all digital systems are guaranteed to suffer this flaw, as all digital systems have finite memory, and therefore will fail as the memory requirement approaches infinity.

Related Research Articles

Linear filters process time-varying input signals to produce output signals, subject to the constraint of linearity. This results from systems composed solely of components classified as having a linear response. Most filters implemented in analog electronics, in digital signal processing, or in mechanical systems are classified as causal, time invariant, and linear signal processing filters.

Interpolation method for constructing new data from known data

In the mathematical field of numerical analysis, interpolation is a method of constructing new data points within the range of a discrete set of known data points.

Kernel may refer to:

Numerical analysis study of algorithms that use numerical approximation for the problems of mathematical analysis

Numerical analysis is the study of algorithms that use numerical approximation for the problems of mathematical analysis. Numerical analysis naturally finds application in all fields of engineering and the physical sciences, but in the 21st century also the life sciences, social sciences, medicine, business and even the arts have adopted elements of scientific computations. As an aspect of mathematics and computer science that generates, analyzes, and implements algorithms, the growth in power and the revolution in computing has raised the use of realistic mathematical models in science and engineering, and complex numerical analysis is required to provide solutions to these more involved models of the world. Ordinary differential equations appear in celestial mechanics ; numerical linear algebra is important for data analysis; stochastic differential equations and Markov chains are essential in simulating living cells for medicine and biology.

Signal processing models and analyzes data representations of physical events

Signal processing is a subfield of mathematics, information and electrical engineering that concerns the analysis, synthesis, and modification of signals, which are broadly defined as functions conveying "information about the behavior or attributes of some phenomenon", such as sound, images, and biological measurements. For example, signal processing techniques are used to improve signal transmission fidelity, storage efficiency, and subjective quality, and to emphasize or detect components of interest in a measured signal.

Filter design is the process of designing a signal processing filter that satisfies a set of requirements, some of which are contradictory. The purpose is to find a realization of the filter that meets each of the requirements to a sufficient degree to make it useful.

In mathematics, deconvolution is an algorithm-based process used to reverse the effects of convolution on recorded data. The concept of deconvolution is widely used in the techniques of signal processing and image processing. Because these techniques are in turn widely used in many scientific and engineering disciplines, deconvolution finds many applications.

In mathematics, an integral transform maps an equation from its original domain into another domain where it might be manipulated and solved much more easily than in the original domain. The solution is then mapped back to the original domain using the inverse of the integral transform.

In signal processing, a finite impulse response (FIR) filter is a filter whose impulse response is of finite duration, because it settles to zero in finite time. This is in contrast to infinite impulse response (IIR) filters, which may have internal feedback and may continue to respond indefinitely.

In numerical analysis, inverse iteration is an iterative eigenvalue algorithm. It allows one to find an approximate eigenvector when an approximation to a corresponding eigenvalue is already known. The method is conceptually similar to the power method. It appears to have originally been developed to compute resonance frequencies in the field of structural mechanics.

Linear programming relaxation linear program formed from a problem in which variables must be 0 or 1 by allowing them to be real numbers between 0 and 1

In mathematics, the relaxation of a (mixed) integer linear program is the problem that arises by removing the integrality constraint of each variable.

hp-FEM is a general version of the finite element method (FEM), a numerical method for solving partial differential equations based on piecewise-polynomial approximations that employs elements of variable size (h) and polynomial degree (p). The origins of hp-FEM date back to the pioneering work of Ivo Babuska et al. who discovered that the finite element method converges exponentially fast when the mesh is refined using a suitable combination of h-refinements (dividing elements into smaller ones) and p-refinements. The exponential convergence makes the method a very attractive choice compared to most other finite element methods which only converge with an algebraic rate. The exponential convergence of the hp-FEM was not only predicted theoretically but also observed by numerous independent researchers.

There are many types of artificial neural networks (ANN).

Neopolarogram

The term neopolarogram refers to mathematical derivatives of polarograms or cyclic voltammograms that in effect deconvolute diffusion and electrochemical kinetics. This is achieved by analog or digital implementations of fractional calculus. The implementation of fractional derivative calculations by means of numerical methods is straight forward. The G1- and the RL0-algorithms are recursive methods to implement a numerical calculation of fractional differintegrals. Yet differintegrals are faster to compute in discrete fourier space using FFT.

In signal processing, multidimensional discrete convolution refers to the mathematical operation between two functions f and g on an n-dimensional lattice that produces a third function, also of n-dimensions. Multidimensional discrete convolution is the discrete analog of the multidimensional convolution of functions on Euclidean space. It is also a special case of convolution on groups when the group is the group of n-tuples of integers.

3D sound refers to the way humans experience sound in their everyday lives. In real life, people are always surrounded by sound. Sounds arrive at our ears from every direction and from varying distances. These and other factors contribute to the three-dimensional aural image humans hear. Scientists and engineers who work with 3D sound work to accurately synthesize the complexity of real-world sounds.