Residence time (statistics)

Last updated

In statistics, the residence time is the average amount of time it takes for a random process to reach a certain boundary value, usually a boundary far from the mean.

Contents

Definition

Suppose y(t) is a real, scalar stochastic process with initial value y(t0) = y0, mean yavg and two critical values {yavgymin, yavg + ymax}, where ymin > 0 and ymax > 0. Define the first passage time of y(t) from within the interval (−ymin, ymax) as

where "inf" is the infimum. This is the smallest time after the initial time t0 that y(t) is equal to one of the critical values forming the boundary of the interval, assuming y0 is within the interval.

Because y(t) proceeds randomly from its initial value to the boundary, τ(y0) is itself a random variable. The mean of τ(y0) is the residence time, [1] [2]

For a Gaussian process and a boundary far from the mean, the residence time equals the inverse of the frequency of exceedance of the smaller critical value, [2]

where the frequency of exceedance N is

 

 

 

 

(1)

σy2 is the variance of the Gaussian distribution,

and Φy(f) is the power spectral density of the Gaussian distribution over a frequency f.

Generalization to multiple dimensions

Suppose that instead of being scalar, y(t) has dimension p, or y(t) ∈ ℝp. Define a domain Ψ ⊂ ℝp that contains yavg and has a smooth boundary ∂Ψ. In this case, define the first passage time of y(t) from within the domain Ψ as

In this case, this infimum is the smallest time at which y(t) is on the boundary of Ψ rather than being equal to one of two discrete values, assuming y0 is within Ψ. The mean of this time is the residence time, [3] [4]

Logarithmic residence time

The logarithmic residence time is a dimensionless variation of the residence time. It is proportional to the natural log of a normalized residence time. Noting the exponential in Equation ( 1 ), the logarithmic residence time of a Gaussian process is defined as [5] [6]

This is closely related to another dimensionless descriptor of this system, the number of standard deviations between the boundary and the mean, min(ymin, ymax)/σy.

In general, the normalization factor N0 can be difficult or impossible to compute, so the dimensionless quantities can be more useful in applications.

See also

Notes

  1. Meerkov & Runolfsson 1987, pp. 1734–1735.
  2. 1 2 Richardson et al. 2014, p. 2027.
  3. Meerkov & Runolfsson 1986, p. 494.
  4. Meerkov & Runolfsson 1987, p. 1734.
  5. Richardson et al. 2014, p. 2028.
  6. Meerkov & Runolfsson 1986, p. 495, an alternate approach to defining the logarithmic residence time and computing N0

Related Research Articles

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Allan variance</span> Measure of frequency stability in clocks and oscillators

The Allan variance (AVAR), also known as two-sample variance, is a measure of frequency stability in clocks, oscillators and amplifiers. It is named after David W. Allan and expressed mathematically as . The Allan deviation (ADEV), also known as sigma-tau, is the square root of the Allan variance, .

<span class="mw-page-title-main">Pink noise</span> Type of signal whose amplitude is inversely proportional to its frequency

Pink noise, 1f noise or fractal noise is a signal or process with a frequency spectrum such that the power spectral density is inversely proportional to the frequency of the signal. In pink noise, each octave interval carries an equal amount of noise energy.

<span class="mw-page-title-main">Fourier transform</span> Mathematical transform that expresses a function of time as a function of frequency

In physics and mathematics, the Fourier transform (FT) is a transform that converts a function into a form that describes the frequencies present in the original function. The output of the transform is a complex-valued function of frequency. The term Fourier transform refers to both this complex-valued function and the mathematical operation. When a distinction needs to be made the Fourier transform is sometimes called the frequency domain representation of the original function. The Fourier transform is analogous to decomposing the sound of a musical chord into terms of the intensity of its constituent pitches.

<span class="mw-page-title-main">Cross-correlation</span> Covariance and correlation

In signal processing, cross-correlation is a measure of similarity of two series as a function of the displacement of one relative to the other. This is also known as a sliding dot product or sliding inner-product. It is commonly used for searching a long signal for a shorter, known feature. It has applications in pattern recognition, single particle analysis, electron tomography, averaging, cryptanalysis, and neurophysiology. The cross-correlation is similar in nature to the convolution of two functions. In an autocorrelation, which is the cross-correlation of a signal with itself, there will always be a peak at a lag of zero, and its size will be the signal energy.

<span class="mw-page-title-main">Stopping time</span> Time at which a random variable stops exhibiting a behavior of interest

In probability theory, in particular in the study of stochastic processes, a stopping time is a specific type of “random time”: a random variable whose value is interpreted as the time at which a given stochastic process exhibits a certain behavior of interest. A stopping time is often defined by a stopping rule, a mechanism for deciding whether to continue or stop a process on the basis of the present position and past events, and which will almost always lead to a decision to stop at some finite time.

<span class="mw-page-title-main">Dirac comb</span> Periodic distribution ("function") of "point-mass" Dirac delta sampling

In mathematics, a Dirac comb is a periodic function with the formula

Nondimensionalization is the partial or full removal of physical dimensions from an equation involving physical quantities by a suitable substitution of variables. This technique can simplify and parameterize problems where measured units are involved. It is closely related to dimensional analysis. In some physical systems, the term scaling is used interchangeably with nondimensionalization, in order to suggest that certain quantities are better measured relative to some appropriate unit. These units refer to quantities intrinsic to the system, rather than units such as SI units. Nondimensionalization is not the same as converting extensive quantities in an equation to intensive quantities, since the latter procedure results in variables that still carry units.

A cyclostationary process is a signal having statistical properties that vary cyclically with time. A cyclostationary process can be viewed as multiple interleaved stationary processes. For example, the maximum daily temperature in New York City can be modeled as a cyclostationary process: the maximum temperature on July 21 is statistically different from the temperature on December 20; however, it is a reasonable approximation that the temperature on December 20 of different years has identical statistics. Thus, we can view the random process composed of daily maximum temperatures as 365 interleaved stationary processes, each of which takes on a new value once per year.

<span class="mw-page-title-main">Gabor transform</span>

The Gabor transform, named after Dennis Gabor, is a special case of the short-time Fourier transform. It is used to determine the sinusoidal frequency and phase content of local sections of a signal as it changes over time. The function to be transformed is first multiplied by a Gaussian function, which can be regarded as a window function, and the resulting function is then transformed with a Fourier transform to derive the time-frequency analysis. The window function means that the signal near the time being analyzed will have higher weight. The Gabor transform of a signal x(t) is defined by this formula:

In mathematics, a local martingale is a type of stochastic process, satisfying the localized version of the martingale property. Every martingale is a local martingale; every bounded local martingale is a martingale; in particular, every local martingale that is bounded from below is a supermartingale, and every local martingale that is bounded from above is a submartingale; however, in general a local martingale is not a martingale, because its expectation can be distorted by large values of small probability. In particular, a driftless diffusion process is a local martingale, but not necessarily a martingale.

In probability theory and statistics, the normal-gamma distribution is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and precision.

In mathematics – specifically, in stochastic analysis – an Itô diffusion is a solution to a specific type of stochastic differential equation. That equation is similar to the Langevin equation used in physics to describe the Brownian motion of a particle subjected to a potential in a viscous fluid. Itô diffusions are named after the Japanese mathematician Kiyosi Itô.

<span class="mw-page-title-main">Quantile regression</span> Statistics concept

Quantile regression is a type of regression analysis used in statistics and econometrics. Whereas the method of least squares estimates the conditional mean of the response variable across values of the predictor variables, quantile regression estimates the conditional median of the response variable. Quantile regression is an extension of linear regression used when the conditions of linear regression are not met.

In linear algebra, the restricted isometry property (RIP) characterizes matrices which are nearly orthonormal, at least when operating on sparse vectors. The concept was introduced by Emmanuel Candès and Terence Tao and is used to prove many theorems in the field of compressed sensing. There are no known large matrices with bounded restricted isometry constants, but many random matrices have been shown to remain bounded. In particular, it has been shown that with exponentially high probability, random Gaussian, Bernoulli, and partial Fourier matrices satisfy the RIP with number of measurements nearly linear in the sparsity level. The current smallest upper bounds for any large rectangular matrices are for those of Gaussian matrices. Web forms to evaluate bounds for the Gaussian ensemble are available at the Edinburgh Compressed Sensing RIC page.

<span class="mw-page-title-main">Modified Allan variance</span>

The modified Allan variance (MVAR), also known as mod σy2(τ), is a variable bandwidth modified variant of Allan variance, a measurement of frequency stability in clocks, oscillators and amplifiers. Its main advantage relative to Allan variance is its ability to separate white phase noise from flicker phase noise.

In time series analysis, the cross-spectrum is used as part of a frequency domain analysis of the cross-correlation or cross-covariance between two time series.

<span class="mw-page-title-main">Exponentially modified Gaussian distribution</span> Describes the sum of independent normal and exponential random variables

In probability theory, an exponentially modified Gaussian distribution describes the sum of independent normal and exponential random variables. An exGaussian random variable Z may be expressed as Z = X + Y, where X and Y are independent, X is Gaussian with mean μ and variance σ2, and Y is exponential of rate λ. It has a characteristic positive skew from the exponential component.

The frequency of exceedance, sometimes called the annual rate of exceedance, is the frequency with which a random process exceeds some critical value. Typically, the critical value is far from the mean. It is usually defined in terms of the number of peaks of the random process that are outside the boundary. It has applications related to predicting extreme events, such as major earthquakes and floods.

In mathematics and theoretical computer science, analysis of Boolean functions is the study of real-valued functions on or from a spectral perspective. The functions studied are often, but not always, Boolean-valued, making them Boolean functions. The area has found many applications in combinatorics, social choice theory, random graphs, and theoretical computer science, especially in hardness of approximation, property testing, and PAC learning.

References