Jump to content

Neyman Type A distribution

From Wikipedia, the free encyclopedia
Neyman Type A
Probability mass function
Probability mass function Neyman Type A
The horizontal axis is the index x, the number of occurrences. The vertical axis we have the probability that the v.a. take the value of x
Cumulative distribution function
Probability Cumulative function Neyman Type A
The horizontal axis is the index x, the number of occurrences. The vertical axis we have the acumulative sum of probabilityes from to
Notation
Parameters
Support x ∈ { 0, 1, 2, ... }
PMF
CDF
Mean
Variance
Skewness
Excess kurtosis
MGF
CF
PGF

In statistics and probability, the Neyman Type A distribution is a discrete probability distribution from the family of Compound Poisson distribution. First of all, to easily understand this distribution we will demonstrate it with the following example explained in Univariate Discret Distributions;[1] we have a statistical model of the distribution of larvae in a unit area of field (in a unit of habitat) by assuming that the variation in the number of clusters of eggs per unit area (per unit of habitat) could be represented by a Poisson distribution with parameter , while the number of larvae developing per cluster of eggs are assumed to have independent Poisson distribution all with the same parameter . If we want to know how many larvae there are, we define a random variable Y as the sum of the number of larvae hatched in each group (given j groups). Therefore, Y = X1 + X2 + ... X j, where X1,...,Xj are independent Poisson variables with parameter and .

History

[edit]

Jerzy Neyman was born in Russia in April 16 of 1894, he was a Polish statistician who spent the first part of his career in Europe. In 1939 he developed the Neyman Type A distribution [1] to describe the distribution of larvae in experimental field plots. Above all, it is used to describe populations based on contagion, e.g., entomology (Beall[1940],[2] Evans[1953][3]), accidents (Creswell i Froggatt [1963]),[4] and bacteriology.

The original derivation of this distribution was on the basis of a biological model and, presumably, it was expected that a good fit to the data would justify the hypothesized model. However, it is now known that it is possible to derive this distribution from different models (William Feller[1943]),[5] and in view of this, Neyman's distribution derive as Compound Poisson distribution. This interpretation makes them suitable for modelling heterogeneous populations and renders them examples of apparent contagion.

Despite this, the difficulties in dealing with Neyman's Type A arise from the fact that its expressions for probabilities are highly complex. Even estimations of parameters through efficient methods, such as maximum likelihood, are tedious and not easy to understand equations.

Definition

[edit]

Probability generating function

[edit]

The probability generating function (pgf) G1(z), which creates N independent Xj random variables, is used to a branching process. Each Xj produces a random number of individuals, where X1, X2,... have the same distribution as X, which is that of X with pgf G2(z). The total number of  individuals is then the random variable,[1]

The p.g.f. of the distribution of SN is :

One of the notations, which is particularly helpful, allows us to use a symbolic representation to refer to  an F1 distribution that has been generalized by an F2 distribution is,

In this instance, it is written as,

Finally, the probability generating function is,

From the generating function of probabilities we can calculate the probability mass function explained below.

Probability mass function

[edit]

Let X1,X2,...Xj be Poisson independent variables. The probability distribution of the random variable Y = X1 +X2+...Xj is the Neyman's Type A distribution with parameters and .

Alternatively,

In order to see how the previous expression develops, we must bear in mind that the probability mass function is calculated from the probability generating function, and use the property of Stirling Numbers. Let's see the development

Another form to estimate the probabilities is with recurring successions,[6]

,

Although its length varies directly with n, this recurrence relation is only employed for numerical computation and is particularly useful for computer applications.

where

  • x = 0, 1, 2, ... , except for probabilities of recurring successions, where x = 1, 2, 3, ...
  • , .
  • x! and j! are the factorials of x and j, respectively.
  • one of the properties of Stirling numbers of the second kind is as follows:[7]

Notation

[edit]

Properties

[edit]

Moment and cumulant generating functions

[edit]

The moment generating function of a random variable X is defined as the expected value of et, as a function of the real parameter t. For an , the moment generating function exists and is equal to

The cumulant generating function is the logarithm of the moment generating function and is equal to [1]

In the following table we can see the moments of the order from 1 to 4

Order Moment Cumulant
1
2
3
4

Skewness

[edit]

The skewness is the third moment centered around the mean divided by the 3/2 power of the standard deviation, and for the distribution is,

Kurtosis

[edit]

The kurtosis is the fourth moment centered around the mean, divided by the square of the variance, and for the distribution is,

The excess kurtosis is just a correction to make the kurtosis of the normal distribution equal to zero, and it is the following,

  • Always , or the distribution has a high acute peak around the mean and fatter tails.

Characteristic function

[edit]

In a discrete distribution the characteristic function of any real-valued random variable is defined as the expected value of , where i is the imaginary unit and t ∈ R

This function is related to the moment generating function via . Hence for this distribution the characteristic function is,

  • Note that the symbol is used to represent the characteristic function.

Cumulative distribution function

[edit]

The cumulative distribution function is,

Other properties

[edit]
  • From a sample of size N, where each random variable Yi comes from a , where Y1, Y2, .., Yn are independent. This gives the MLE estimator as,[9]
where is the poblational mean of
  • Between the two earlier expressions We are able to parametrize using and ,

Parameter estimation

[edit]

Method of moments

[edit]

The mean and the variance of the NA() are and , respectively. So we have these two equations,[10]

  • and are the mostral variance and mean respectively.

Solving these two equations we get the moment estimators and of and .

Maximum likelihood

[edit]

Calculating the maximum likelihood estimator of and involves multiplying all the probabilities in the probability mass function to obtain the expression .

When we apply the parameterization adjustment defined in "Other Properties," we get . We may define the Maximum likelihood estimation based on a single parameter if we estimate the as the (sample mean) given a sample X of size N. We can see it below.

  • To estimate the probabilities, we will use the p.m.f. of recurring successions, so that the calculation is less complex.

Testing Poisson assumption

[edit]

When is used to simulate a data sample it is important to see if the Poisson distribution fits the data well. For this, the following Hypothesis test is used:

Likelihood-ratio test

[edit]

The likelihood-ratio test statistic for is,

Where likelihood is the log-likelihood function. W does not have an asymptotic distribution as expected under the null hypothesis since d = 1 is at the parameter domain's edge. In the asymptotic distribution of W, it can be demonstrated that the constant 0 and have a 50:50 mixture. For this mixture, the upper-tail percentage points are the same as the upper-tail percentage points for a

[edit]

The poisson distribution (on { 0, 1, 2, 3, ... }) is a special case of the Neyman Type A distribution, with

From moments of order 1 and 2 we can write the population mean and variance based on the parameters and .

In the dispersion index d we observe that by substituting for the parametrized equation of order 1 and for the one of order 2, we obtain . Our variable Y is therefore distributed as a Poisson of parameter when d approaches to 1.

Then we have that,

Applications

[edit]

Usage History

[edit]

When the Neyman's type A species' reproduction results in clusters, a distribution has been used to characterize the dispersion of plants. This typically occurs when a species develops from offspring of parent plants or from seeds that fall close to the parent plant. However, Archibald (1948][11] observed that there is insufficient data to infer the kind of reproduction from the type of fitted distribution. While Neyman type A produced positive results for plant distributions, Evans (1953][3] showed that Negative binomial distribution produced positive results for insect distributions. Neyman type A distributions have also been studied in the context of ecology, and the results that unless the plant clusters are so compact as to not lie across the edge of the square used to pick sample locations, the distribution is unlikely to be applicable to plant populations. The compactness of the clusters is a hidden assumption in Neyman's original derivation of the distribution, according to Skellam (1958).[12] The results were shown to be significantly impacted by the square size selection.

In the context of bus driver accidents, Cresswell and Froggatt (1963) [4] derived the Neyman type A based on the following hypotheses:

  1. Each driver is susceptible to "spells," the number of which is Poissonian for any given amount of time, with the same parameter for all drivers.
  2. A driver's performance during a spell is poor, and he is likely to experience a Poissonian number of collisions, with the same parameter for all drivers.
  3. Each driver behaves independently.
  4. No accidents can occur outside of a spell.

These assumptions lead to a Neyman type A distribution via the and model. In contrast to their "short distribution," Cresswell and Froggatt called this one the "long distribution" because of its lengthy tail. According to Irwin(1964), a type A distribution can also be obtained by assuming that various drivers have different levels of proneness, or K. with probability:

  • taking values

and that a driver with proneness k has X accidents where:

This is the model with mixing over the values taken by K.

Distribution was also suggested in the application for the grouping of minority tents for food from 1965 to 1969. In this regard, it was predicted that only the clustering rates or the average number of entities per grouping needed to be approximated, rather than adjusting distributions on d very big data bases.

  • The reference of the usage history.[1]

Calculating Neyman Type A probabilities in R

[edit]
  • The code below simulates 5000 instances of ,
 rNeymanA <- function(n,lambda, phi){
   r <- numeric()
   for (j in 1:n) {
      k = rpois(1,lambda)
      r[j] <- sum(rpois(k,phi)) 
   }
   return(r)
 }
  • The mass function of recurring probabilities is implemented in R in order to estimate theoretical probabilities; we can see it below,
dNeyman.rec <- function(x, lambda, phi){
 p <- numeric()
 p[1]<- exp(-lambda + lambda*exp(-phi))
 c <- lambda*phi*exp(-phi)
 if(x == 0){
   return(p[1])
 }
 else{
   for (i in 1:x) {
     suma = 0
     for (r in 0:(i-1)) {
       suma = suma + (phi^r)/(factorial(r))*p[i-r]
     }
     p[i+1] = (c/(i))*suma # +1 per l'R
   }
   res <- p[i+1]
   return(res)
 }
}

We compare results between the relative frequencies obtained by the simulation and the probabilities computed by the p.m.f. . Given two values for the parameters and . It is displayed in the following table,

X 0 1 2 3 4 5 6 7 8
Estimated .271 .207 .187 .141 .082 .052 .028 .016 .008
Teoric .282 .207 .180 .129 .084 .051 .029 .016 .008

References

[edit]
  1. ^ a b c d e Johnson, N.L; Kemp, A.W; Kotz, S (2005). Univariate Discret Distributions. Canada: Wiley-Intersciencie. pp. 407–409. ISBN 0-471-27246-9.
  2. ^ Beall, G. (1940). "The fit and significance of contagious distributions when applied to observations on larval insects". Ecology. 21 (4): 460–474. Bibcode:1940Ecol...21..460B. doi:10.2307/1930285. JSTOR 1930285.
  3. ^ a b Evans, D.A. (1953). "Experimental evidence concerning contagious distributions in ecology". Biometrika. 40 (1–2): 186–211. doi:10.1093/biomet/40.1-2.186.
  4. ^ a b Creswell, W.L.; Froggatt, P. (1963). The Causation of Bus Driver Accidents. London: Oxford University Press.
  5. ^ Feller, W. (1943). "On a general class of "contagious" distributions". Annals of Mathematical Statistics. 14 (4): 389. doi:10.1214/aoms/1177731359.
  6. ^ Johnson, N.L; Kemp, A.W; Kotz, S (2005). Univariate Discret Distributions. Canada: Wiley-Intersciencie. p. 404.
  7. ^ Johnson, N.L; Kemp, A.W; Kotz, S (2005). "1.1.3". Univariate Discret Distributions. Canada: Wiley-Intersciencie.
  8. ^ "Índice de dispersión (Varianza a Razón media)". Statologos. April 18, 2022.
  9. ^ Shenton, L.R; Bowman, K.O (1977). "9.6.1". Maximum Likelihood Estimation in Small Samples. London: Griffin.
  10. ^ Shenton, L.R (1949). "On the efficiency of the method of moments and Neyman Type A Distribution". Biometrika. 36 (3–4): 450–454. doi:10.1093/biomet/36.3-4.450. JSTOR 2332680. PMID 15402077.
  11. ^ Archibald, E.E.A. Plant populations I: A new application of Neyman's contagious distribution, Annals of Botany. London: Oxford University Press. pp. 221–235.
  12. ^ Skellam, J.G. (1958). "On the derivation and applicability of Neyman's type A distribution". Biometrika. 45 (1–2): 32–36. doi:10.1093/biomet/45.1-2.32.