STAT 400

Mon. February 24th, 2020


Poisson Distribution

Definition. A discrete random variable XX follows the Poisson distribution with parameter μ\mu (μ>0\mu >0) if its probabilty mass function is P(X=k)=p(k;μ)=μkk!eμP(X=k)=p(k;\mu)=\frac{\mu^k}{k!}e^{-\mu} for k=0,1,2...k=0,1,2...

The cumulative distribution function for XX has no closed form; you should use a Poisson table to calculate it.


Observation. Recall that the Taylor series expansion of eμe^\mu is eμ=1+μ+μ22!+μ33!+...=k=0μkk!e^\mu=1+\mu +\frac{\mu^2}{2!}+\frac{\mu^3}{3!}+...=\sum_{k=0}^\infty \frac{\mu^k}{k!} Using this fact, we see that the sum of all possible values of p(k;μ)p(k;\mu) is k=0p(k;μ)=k=0μkk!eμ=eμeμ=1,\sum_{k=0}^\infty p(k;\mu) =\sum_{k=0}^\infty \frac{\mu^k}{k!}e^{-\mu}=e^\mu e^{-\mu}=1, which is what we’d expect for a useful distribution.


Proposition. Suppose XX follows bin(n,p)\text{bin}(n,p). As nn\rightarrow \infin, p0p\rightarrow 0, and npμnp\rightarrow \mu, then XX converges towards the Poisson distribution Pois(μ)\text{Pois}(\mu).

(Rule of thumb: When n>50n>50 and np<5np<5, we can use the above proposition to approximate Pois(μ)\text{Pois}(\mu) with bin(n,p)\text{bin}(n,p) and vice-versa.)

The Poisson distribution can be used to model the number of random visits/arrivals that will occur in a fixed period of time.

Ex. 100 people pass by a store in 10 minutes. The probability that any one enters the store is 0.02. If XX is the number of people that entered, then XX follows bin(100,0.02)\text{bin}(100,0.02).
Following our rule of thumb, we can approximate the probability mass function of XX with Pois(μ)=Pois(np)=Pois(2)\text{Pois}(\mu)=\text{Pois}(np)=\text{Pois}(2).


Expected Value and Variance

Suppose XX follows Pois(μ)\text{Pois}(\mu). Then E(X)=μE(X)=\mu and V(X)=μV(X)=\mu. This interesting property is another reason we study the Poisson distribution.


Sum of Poisson Distributions

Suppose a collection of random variables XiX_i all follow Pois(μi)\text{Pois}(\mu_i). Think of them as representing the number of visits that occur in a sequence of time periods – they are independent.

If we sum all XiX_i to produce a new random variable YY, then YY also follows Pois(μ)\text{Pois}(\mu), where μ\mu is the sum of all μi\mu_i. (Binomial distributions also have this property.)


Poisson Process

Suppose X1X_1 follows Pois(t1μ)\text{Pois}(t_1\mu) and represents the number of visits from time 0 to time t1t_1, and X2X_2 follows Pois(t2μ)\text{Pois}(t_2\mu) and represents the number of visits from time t1t_1 to time t1+t2t_1+t_2. (Note that t1t_1 and t2t_2 are durations of time, not points in time.)

Then XX, the number of visits from time 0 to time t1+t2t_1+t_2, is simply X1+X2X_1+X_2 and follows Pois((t1+t2)μ)\text{Pois}((t_1+t_2)\mu), from the sum property described above.


Definition. Given parameter μ\mu, suppose visits occur at times T1T2T3...T_1\le T_2\le T_3\le ...
Let X(s,t)X(s,t) be the number of visits from time ss to time tt, with 0<s<t0< s < t. In other words, X(s,t)=N{i:Ti[s,t]}X(s,t)=N\{i: T_i\in [s,t]\}.

If X(s,t)X(s,t) follows Pois((ts)μ)\text{Pois}((t-s)\mu), then we call the collection of {X(s,t):0<s<t}\{X(s,t):0< s< t\} a Poisson process. Here, μ\mu is the parameter of the Poisson process, also called the rate of the process.


Properties

From the observations we’ve made before, we see that X(0,t1)+X(t1, t1+t2)=X(0, t1+t2)X(0,t_1)+X(t_1,\ t_1+t_2)=X(0,\ t_1+t_2); the sum of visits of two consecutive intervals of time is the same as visits made over the union of those intervals, which makes sense intuitively. This can also be expanded to sum over any number of intervals.


Expected Value and Variance of Sums of Discrete Random Variables

If X1,X2,...XnX_1,X_2,...X_n are arbitrary discrete random variables, then E(X1+X2+...+Xn)=E(X1)+E(X2)+...+E(Xn)E(X_1+X_2+...+X_n)=E(X_1)+E(X_2)+...+E(X_n) since EE is linear.

In addition, if these variables are all independent and identically distributed, then V(X1+X2+...+Xn)=nV(X1).V(X_1+X_2+...+X_n)=nV(X_1).