Discrete Distributions 2
4. Poisson Distribution
4.1 Definition
A discrete random variable is said to follow the poisson distribution P(λ) with λ∈R+ if:
∀n∈N,P(X=n)=n!λne−λ 4.2 Significance
In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known constant mean rate and independently of the time since the last event.[1] It is named after French mathematician Siméon Denis Poisson (/ˈpwɑːsɒn/; French pronunciation: [pwasɔ̃]). The Poisson distribution can also be used for the number of events in other specified interval types such as distance, area, or volume.
For instance, a call center receives an average of 180 calls per hour, 24 hours a day. The calls are independent; receiving one does not change the probability of when the next one will arrive. The number of calls received during any minute has a Poisson probability distribution with mean 3: the most likely numbers are 2 and 3 but 1 and 4 are also likely and there is a small probability of it being as low as zero and a very small probability it could be 10.
Another example is the number of decay events that occur from a radioactive source during a defined observation period.
4.3 Moments
4.3.1 Raw Moments
∀n∈N∗,E[Xn]=m∈N∑m!mnλme−λ=m∈N∗∑(m−1)!mn−1λme−λ=λm∈N∑m!(m+1)n−1λme−λ=λm∈N∑s=0∑n−1(sn−1)m!msλme−λ=λs=0∑n−1(sn−1)m∈N∑m!msλme−λ=λs=0∑n−1(sn−1)E[Xs] In particular:
E[X]=λE[X0]=λE[1]=λ 4.3.2 Central Moments
We will start by the variance:
E[X2]⟹V[X]=λs=0∑1(s1)E[Xs]=λE[X0]+λE[X1]=λ+λ2=E[X2]−E[X]2=λ 5. Hyper-geometric Distribution
5.1 Prelude
For a random variable X and event E.
We will denote by X[E] the random variable X knowing that E occured
5.2 Definition
Let N,K,n∈N/K≤N and n≤N
We say that a random variable X follows the hyper-geometric distribution H(n,N,K) with parameters N,K,n if:
∃X1,…,XN∼B(p) i.i.d /X=Sn[SN=K]with Sk=i=1∑kXi∀k∈{1,…,N} 5.2 Significance
In probability theory and statistics, the hyper-geometric distribution is a discrete probability distribution that describes the probability of k successes (random draws for which the object drawn has a specified feature) in n draws, without replacement, from a finite population of size N that contains exactly K objects with that feature, wherein each draw is either a success or a failure.
In contrast, the binomial distribution describes the probability of successes in n draws with replacement.
5.3 Probability Mass function
We have Sn=∑i=1nXi∼B(n,p) and SN−Sn=∑i=n+1NXi∼B(N−n,p).
In fact, Sn and SN−Sn are independent.
With that we have:
∀k∈{0,…,K},P(X=k)=P(Sn=k∣SN=K)=P(SN=K)P(Sn=k ∧ SN=K)=P(SN=K)P(Sn=k ∧ SN−Sn=K−k)=P(SN=K)P(Sn=k)⋅P(SN−Sn=K−k)=(KN)pK(1−p)N−K(kn)pk(1−p)n−k×(K−kN−n)pK−k(1−p)N−n−K+k=(KN)(kn)×(K−kN−n)=k!(n−k)!(K−k)!(N−n−K+k)!N!n!(N−n)!K!(N−K)!=k!(K−k)!K!⋅(n−k)!(N−n−n+k)!(N−K)!⋅N!n!(N−n)!=(nN)(kK)⋅(n−kN−K) 5.4 Moments
5.4.1 Raw Moments
We will start by the expected value E[X]:
E[X]=E[Sn∣SN=K]=k=1∑nE[Xk∣SN=K]=k=1∑n P(Xk=1∣SN=K)=k=1∑nP(SN=K)∑k=1n P(Xk=1∧SN=K)=k=1∑nP(SN=K)∑k=1n P(Xk=1∧SN−Xk=K−1)=k=1∑nP(SN=K)P(Xk=1)⋅P(SN−Xk=K−1)as SN−Xk and Xk are independent=k=1∑n(KN)pK(1−P)N−Kp⋅(K−1N−1)pK−1(1−p)N−K=k=1∑n(KN)(K−1N−1)=n⋅K!(N−K)!N!(N−1)!K!(N−K)!=nKN 5.4.2 Central Moments
For convenience, we will denote Yi=Xi[Sn=K]
We will start by the variance V[X]:
V[X]∀i=j,Cov(Yi,Yj)⟹V[X]=Cov(Sn[SN=K],Sn[SN=K])=Cov((i=1∑nXi)[SN=K],(j=1∑nXj)[SN=K])=Cov(i=1∑nXi[SN=K],j=1∑nXj[SN=K])=i=1∑nV[Yi]+21≤i<j≤n∑Cov(Yi,Yj)=E[Yi×Yj]−E[Yi]×E[Yj]=P[(Yi×Yj)=1]−N2K2=P((Yi×Yj)=1∣Yj=1)×P(Yj=1)−N2K2=P(Yi=1∣Yj=1)×P(Yj=1)−N2K2=P(Yi=1∣Yj=1∧SN=K)×NK−N2K2=P(Xi=1∣Xj=1∧SN=K)×NK−N2K2=P(Xj=1∧Sn=K)P(Xi=1∧Xj=1∧SN=K)×NK−N2K2=P(Xj=1∧SN−Xj=K−1)P(Xi=1∧Xj=1∧SN−Xi−Xj=K−2)×NK−N2K2=P(Xj=1)⋅P(SN−Xj=K−1)P(Xi=1)⋅P(Xj=1)⋅P(SN−Xi−Xj=K−2)×NK−N2K2=p⋅(K−1N−1)pK−1(1−p)N−Kp⋅p⋅(K−2N−2)pK−2(1−p)N−K×NK−N2K2=N−1K−1×NK−N2K2=NK(N−1K−1−NK)=i=1∑nV[Yi]+21≤i<j≤n∑Cov(Yi,Yj)=nNKNN−K+2×2n(n−1)⋅NK(N−1K−1−NK)=NK(nNN−K+n(n−1)⋅(N−1K−1−NK))=nNK(NN−K+(n−1)⋅(N−1K−1−NK))=nNK(N(N−1)(N−K)(N−1)+N(n−1)(K−1)−K(n−1)(N−1))=nNK(N(N−1)N2−KN−N+K+nNK−NK−nN+N−nNK+KN+nK−K)=nNK(N(N−1)N2−NK−nN+nK)=nNK(N(N−1)N(N−K)−n(N−K))=nNK⋅NN−K⋅N−1N−n