Neyman Type A distribution

Last updated
Neyman Type A
Probability mass function
Tfg1plotdensity.png
The horizontal axis is the index x, the number of occurrences. The vertical axis we have the probability that the v.a. take the value of x
Cumulative distribution function
Tfg2plotdistribution.png
The horizontal axis is the index x, the number of occurrences. The vertical axis we have the acumulative sum of probabilityes from to
Notation
Parameters
Support x ∈ { 0, 1, 2, ... }
PMF
CDF
Mean
Variance
Skewness
Ex. kurtosis
MGF
CF
PGF

In statistics and probability, the Neyman Type A distribution is a discrete probability distribution from the family of Compound Poisson distribution. First of all, to easily understand this distribution we will demonstrate it with the following example explained in Univariate Discret Distributions; [1] we have a statistical model of the distribution of larvae in a unit area of field (in a unit of habitat) by assuming that the variation in the number of clusters of eggs per unit area (per unit of habitat) could be represented by a Poisson distribution with parameter , while the number of larvae developing per cluster of eggs are assumed to have independent Poisson distribution all with the same parameter . If we want to know how many larvae there are, we define a random variable Y as the sum of the number of larvae hatched in each group (given j groups). Therefore, Y = X1 + X2 + ... Xj, where X1,...,Xj are independent Poisson variables with parameter and .

Contents

History

Jerzy Neyman was born in Russia in April 16 of 1894, he was a Polish statistician who spent the first part of his career in Europe. In 1939 he developed the Neyman Type A distribution [1] to describe the distribution of larvae in experimental field plots. Above all, it is used to describe populations based on contagion, e.g., entomology (Beall[1940], [2] Evans[1953] [3] ), accidents (Creswell i Froggatt [1963]), [4] and bacteriology.

The original derivation of this distribution was on the basis of a biological model and, presumably, it was expected that a good fit to the data would justify the hypothesized model. However, it is now known that it is possible to derive this distribution from different models (William Feller[1943]), [5] and in view of this, Neyman's distribution derive as Compound Poisson distribution. This interpretation makes them suitable for modelling heterogeneous populations and renders them examples of apparent contagion.

Despite this, the difficulties in dealing with Neyman's Type A arise from the fact that its expressions for probabilities are highly complex. Even estimations of parameters through efficient methods, such as maximum likelihood, are tedious and not easy to understand equations.

Definition

Probability generating function

The probability generating function (pgf) G1(z), which creates N independent Xj random variables, is used to a branching process. Each Xj produces a random number of individuals, where X1, X2,... have the same distribution as X, which is that of X with pgf G2(z). The total number of  individuals is then the random variable, [1]

The p.g.f. of the distribution of SN is :

One of the notations, which is particularly helpful, allows us to use a symbolic representation to refer to  an F1 distribution that has been generalized by an F2 distribution is,

In this instance, it is written as,

Finally, the probability generating function is,

From the generating function of probabilities we can calculate the probability mass function explained below.

Probability mass function

Let X1,X2,...Xj be Poisson independent variables. The probability distribution of the random variable Y = X1 +X2+...Xj is the Neyman's Type A distribution with parameters and .

Alternatively,

In order to see how the previous expression develops, we must bear in mind that the probability mass function is calculated from the probability generating function, and use the property of Stirling Numbers. Let's see the development

Another form to estimate the probabilities is with recurring successions, [6]

,

Although its length varies directly with n, this recurrence relation is only employed for numerical computation and is particularly useful for computer applications.

where

Notation

Properties

Moment and cumulant generating functions

The moment generating function of a random variable X is defined as the expected value of et, as a function of the real parameter t. For an , the moment generating function exists and is equal to

The cumulant generating function is the logarithm of the moment generating function and is equal to [1]

In the following table we can see the moments of the order from 1 to 4

OrderMomentCumulant
1
2
3
4

Skewness

The skewness is the third moment centered around the mean divided by the 3/2 power of the standard deviation, and for the distribution is,

Kurtosis

The kurtosis is the fourth moment centered around the mean, divided by the square of the variance, and for the distribution is,

The excess kurtosis is just a correction to make the kurtosis of the normal distribution equal to zero, and it is the following,

Characteristic function

In a discrete distribution the characteristic function of any real-valued random variable is defined as the expected value of , where i is the imaginary unit and t  R

This function is related to the moment generating function via . Hence for this distribution the characteristic function is,

Cumulative distribution function

The cumulative distribution function is,

Other properties

where is the poblational mean of

Parameter estimation

Method of moments

The mean and the variance of the NA() are and , respectively. So we have these two equations, [10]

Solving these two equations we get the moment estimators and of and .

Maximum likelihood

Calculating the maximum likelihood estimator of and involves multiplying all the probabilities in the probability mass function to obtain the expression .

When we apply the parameterization adjustment defined in "Other Properties," we get . We may define the Maximum likelihood estimation based on a single parameter if we estimate the as the (sample mean) given a sample X of size N. We can see it below.

Testing Poisson assumption

When is used to simulate a data sample it is important to see if the Poisson distribution fits the data well. For this, the following Hypothesis test is used:

Likelihood-ratio test

The likelihood-ratio test statistic for is,

Where likelihood is the log-likelihood function. W does not have an asymptotic distribution as expected under the null hypothesis since d = 1 is at the parameter domain's edge. In the asymptotic distribution of W, it can be demonstrated that the constant 0 and have a 50:50 mixture. For this mixture, the upper-tail percentage points are the same as the upper-tail percentage points for a

The poisson distribution (on { 0, 1, 2, 3, ... }) is a special case of the Neyman Type A distribution, with

From moments of order 1 and 2 we can write the population mean and variance based on the parameters and .

In the dispersion index d we observe that by substituting for the parametrized equation of order 1 and for the one of order 2, we obtain . Our variable Y is therefore distributed as a Poisson of parameter when d approaches to 1.

Then we have that,

Applications

Usage History

When the Neyman's type A species' reproduction results in clusters, a distribution has been used to characterize the dispersion of plants. This typically occurs when a species develops from offspring of parent plants or from seeds that fall close to the parent plant. However, Archibald (1948] [11] observed that there is insufficient data to infer the kind of reproduction from the type of fitted distribution. While Neyman type A produced positive results for plant distributions, Evans (1953] [3] showed that Negative binomial distribution produced positive results for insect distributions. Neyman type A distributions have also been studied in the context of ecology, and the results that unless the plant clusters are so compact as to not lie across the edge of the square used to pick sample locations, the distribution is unlikely to be applicable to plant populations. The compactness of the clusters is a hidden assumption in Neyman's original derivation of the distribution, according to Skellam (1958). [12] The results were shown to be significantly impacted by the square size selection.

In the context of bus driver accidents, Cresswell and Froggatt (1963) [4] derived the Neyman type A based on the following hypotheses:

  1. Each driver is susceptible to "spells," the number of which is Poissonian for any given amount of time, with the same parameter for all drivers.
  2. A driver's performance during a spell is poor, and he is likely to experience a Poissonian number of collisions, with the same parameter for all drivers.
  3. Each driver behaves independently.
  4. No accidents can occur outside of a spell.

These assumptions lead to a Neyman type A distribution via the and model. In contrast to their "short distribution," Cresswell and Froggatt called this one the "long distribution" because of its lengthy tail. According to Irwin(1964), a type A distribution can also be obtained by assuming that various drivers have different levels of proneness, or K. with probability:

and that a driver with proneness k has X accidents where:

This is the model with mixing over the values taken by K.

Distribution was also suggested in the application for the grouping of minority tents for food from 1965 to 1969. In this regard, it was predicted that only the clustering rates or the average number of entities per grouping needed to be approximated, rather than adjusting distributions on d very big data bases.

Calculating Neyman Type A probabilities in R

 rNeymanA <- function(n,lambda, phi){    r <- numeric()    for (j in 1:n) {       k = rpois(1,lambda)       r[j] <- sum(rpois(k,phi))     }    return(r)  }
dNeyman.rec <- function(x, lambda, phi){  p <- numeric()  p[1]<- exp(-lambda + lambda*exp(-phi))  c <- lambda*phi*exp(-phi)  if(x == 0){    return(p[1])  }  else{    for (i in 1:x) {      suma = 0      for (r in 0:(i-1)) {        suma = suma + (phi^r)/(factorial(r))*p[i-r]      }      p[i+1] = (c/(i))*suma # +1 per l'R    }    res <- p[i+1]    return(res)  } }

We compare results between the relative frequencies obtained by the simulation and the probabilities computed by the p.m.f. . Given two values for the parameters and . It is displayed in the following table,

X012345678
Estimated .271.207.187.141.082.052.028.016.008
Teoric .282.207.180.129.084.051.029.016.008

Related Research Articles

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Exponential distribution</span> Probability distribution

In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average rate; the distance parameter could be any meaningful mono-dimensional measure of the process, such as time between production errors, or length along a roll of fabric in the weaving manufacturing process. It is a particular case of the gamma distribution. It is the continuous analogue of the geometric distribution, and it has the key property of being memoryless. In addition to being used for the analysis of Poisson point processes it is found in various other contexts.

<span class="mw-page-title-main">Jensen's inequality</span> Theorem of convex functions

In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906, building on an earlier proof of the same inequality for doubly-differentiable functions by Otto Hölder in 1889. Given its generality, the inequality appears in many forms depending on the context, some of which are presented below. In its simplest form the inequality states that the convex transformation of a mean is less than or equal to the mean applied after convex transformation; it is a simple corollary that the opposite is true of concave transformations.

In probability theory, the Borel–Kolmogorov paradox is a paradox relating to conditional probability with respect to an event of probability zero. It is named after Émile Borel and Andrey Kolmogorov.

The Gram–Charlier A series, and the Edgeworth series are series that approximate a probability distribution in terms of its cumulants. The series are the same; but, the arrangement of terms differ. The key idea of these expansions is to write the characteristic function of the distribution whose probability density function f is to be approximated in terms of the characteristic function of a distribution with known and suitable properties, and to recover f through the inverse Fourier transform.

In probability theory, a compound Poisson distribution is the probability distribution of the sum of a number of independent identically-distributed random variables, where the number of terms to be added is itself a Poisson-distributed variable. The result can be either a continuous or a discrete distribution.

<span class="mw-page-title-main">Laplace distribution</span> Probability distribution

In probability theory and statistics, the Laplace distribution is a continuous probability distribution named after Pierre-Simon Laplace. It is also sometimes called the double exponential distribution, because it can be thought of as two exponential distributions spliced together along the abscissa, although the term is also sometimes used to refer to the Gumbel distribution. The difference between two independent identically distributed exponential random variables is governed by a Laplace distribution, as is a Brownian motion evaluated at an exponentially distributed random time. Increments of Laplace motion or a variance gamma process evaluated over the time scale also have a Laplace distribution.

Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:

  1. To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
  2. To derive a lower bound for the marginal likelihood of the observed data. This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data.

In Bayesian probability, the Jeffreys prior, named after Sir Harold Jeffreys, is a non-informative prior distribution for a parameter space; its density function is proportional to the square root of the determinant of the Fisher information matrix:

<span class="mw-page-title-main">Noncentral chi-squared distribution</span> Noncentral generalization of the chi-squared distribution

In probability theory and statistics, the noncentral chi-squared distribution is a noncentral generalization of the chi-squared distribution. It often arises in the power analysis of statistical tests in which the null distribution is a chi-squared distribution; important examples of such tests are the likelihood-ratio tests.

In theoretical physics, a source field is a background field coupled to the original field as

<span class="mw-page-title-main">Inverse Gaussian distribution</span> Family of continuous probability distributions

In probability theory, the inverse Gaussian distribution is a two-parameter family of continuous probability distributions with support on (0,∞).

Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

In mathematics, the spectral theory of ordinary differential equations is the part of spectral theory concerned with the determination of the spectrum and eigenfunction expansion associated with a linear ordinary differential equation. In his dissertation, Hermann Weyl generalized the classical Sturm–Liouville theory on a finite closed interval to second order differential operators with singularities at the endpoints of the interval, possibly semi-infinite or infinite. Unlike the classical case, the spectrum may no longer consist of just a countable set of eigenvalues, but may also contain a continuous part. In this case the eigenfunction expansion involves an integral over the continuous part with respect to a spectral measure, given by the Titchmarsh–Kodaira formula. The theory was put in its final simplified form for singular differential equations of even degree by Kodaira and others, using von Neumann's spectral theorem. It has had important applications in quantum mechanics, operator theory and harmonic analysis on semisimple Lie groups.

Financial models with long-tailed distributions and volatility clustering have been introduced to overcome problems with the realism of classical financial models. These classical models of financial time series typically assume homoskedasticity and normality cannot explain stylized phenomena such as skewness, heavy tails, and volatility clustering of the empirical asset returns in finance. In 1963, Benoit Mandelbrot first used the stable distribution to model the empirical distributions which have the skewness and heavy-tail property. Since -stable distributions have infinite -th moments for all , the tempered stable processes have been proposed for overcoming this limitation of the stable distribution.

<span class="mw-page-title-main">Poisson distribution</span> Discrete probability distribution

In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event. It can also be used for the number of events in other types of intervals than time, and in dimension greater than 1.

A geometric stable distribution or geo-stable distribution is a type of leptokurtic probability distribution. Geometric stable distributions were introduced in Klebanov, L. B., Maniya, G. M., and Melamed, I. A. (1985). A problem of Zolotarev and analogs of infinitely divisible and stable distributions in a scheme for summing a random number of random variables. These distributions are analogues for stable distributions for the case when the number of summands is random, independent of the distribution of summand, and having geometric distribution. The geometric stable distribution may be symmetric or asymmetric. A symmetric geometric stable distribution is also referred to as a Linnik distribution. The Laplace distribution and asymmetric Laplace distribution are special cases of the geometric stable distribution. The Mittag-Leffler distribution is also a special case of a geometric stable distribution.

In the theory of renewal processes, a part of the mathematical theory of probability, the residual time or the forward recurrence time is the time between any given time and the next epoch of the renewal process under consideration. In the context of random walks, it is also known as overshoot. Another way to phrase residual time is "how much more time is there to wait?".

In statistics, the variance function is a smooth function that depicts the variance of a random quantity as a function of its mean. The variance function is a measure of heteroscedasticity and plays a large role in many settings of statistical modelling. It is a main ingredient in the generalized linear model framework and a tool used in non-parametric regression, semiparametric regression and functional data analysis. In parametric modeling, variance functions take on a parametric form and explicitly describe the relationship between the variance and the mean of a random quantity. In a non-parametric setting, the variance function is assumed to be a smooth function.

References

  1. 1 2 3 4 5 Johnson, N.L; Kemp, A.W; Kotz, S (2005). Univariate Discret Distributions. Canada: Wiley-Intersciencie. pp. 407–409. ISBN   0-471-27246-9.
  2. Beall, G. (1940). "The fit and significance of contagious distributions when applied to observations on larval insects". Ecology. 21 (4): 460–474. Bibcode:1940Ecol...21..460B. doi:10.2307/1930285. JSTOR   1930285.
  3. 1 2 Evans, D.A. (1953). "Experimental evidence concerning contagious distributions in ecology". Biometrika. 40 (1–2): 186–211. doi:10.1093/biomet/40.1-2.186.
  4. 1 2 Creswell, W.L.; Froggatt, P. (1963). The Causation of Bus Driver Accidents. London: Oxford University Press.
  5. Feller, W. (1943). "On a general class of "contagious" distributions". Annals of Mathematical Statistics. 14 (4): 389. doi:10.1214/aoms/1177731359.
  6. Johnson, N.L; Kemp, A.W; Kotz, S (2005). Univariate Discret Distributions. Canada: Wiley-Intersciencie. p. 404.
  7. Johnson, N.L; Kemp, A.W; Kotz, S (2005). "1.1.3". Univariate Discret Distributions. Canada: Wiley-Intersciencie.
  8. "Índice de dispersión (Varianza a Razón media)". Statologos. April 18, 2022.
  9. Shenton, L.R; Bowman, K.O (1977). "9.6.1". Maximum Likelihood Estimation in Small Samples. London: Griffin.
  10. Shenton, L.R (1949). "On the efficiency of the method of moments and Neyman Type A Distribution". Biometrika. 36: 450–454. doi:10.1093/biomet/36.3-4.450. JSTOR   2332680.
  11. Archibald, E.E.A. Plant populations I: A new application of Neyman's contagious distribution, Annals of Botany. London: Oxford University Press. pp. 221–235.
  12. Skellam, J.G. (1958). "On the derivation and applicability of Neyman's type A distribution". Biometrika. 45 (1–2): 32–36. doi:10.1093/biomet/45.1-2.32.