The Poisson Process

The Poisson process counts independent, exponentially-spaced arrivals in a fixed interval.

Posted on Mathematics Simulation Statistics

Previously, we modeled the time between independent events using the exponential distribution. We gave the example of customers arriving at a coffee shop at a rate of λ=3\lambda = 3 per hour (where the tick is one hour). The time between arrivals follows Exp(3)\text{Exp}(3). This gives us an expected gap of 20 minutes between customers.

Note: If the tick were 1 minute, instead of 1 hour, λ=0.05\lambda = 0.05. The tick is important!

Now we flip the question. Instead of asking, “How long until the next customer?”, we ask: “How many customers arrive in a given window of time?” If we watch the shop for an entire 8-hour shift, how many customers should we expect?

The expected value is obvious. An average of 3 customers per hour. 8 hours. We should expect 24 customers in 8 hours. But what about every other possibility?

The answer is the Poisson distribution.

From Exponential to Poisson

The derivation starts with the same setup. Events arrive at a constant average rate λ\lambda. We want to count how many events occur in a fixed interval of length tt.

Step 1: Partition the interval. Slice [0,t][0, t] into nn equal subintervals, each of length tn\frac{t}{n}.

Step 2: Probability per subinterval. For large nn, each subinterval is so short that at most one event can fall in it. The probability of exactly one event in a subinterval of length tn\frac{t}{n} is approximately:

P(one event)λtnP(\text{one event}) \approx \lambda \cdot \frac{t}{n}

Step 3: Recognize the binomial. Each subinterval either contains an event or it doesn’t, independently of the others. That’s a classic binomial setup: nn independent trials, each with success probability p=λtnp = \frac{\lambda t}{n}. The total number of events across all nn subintervals follows:

XBinomial ⁣(n,p)=Binomial ⁣(n,λtn)X \sim \text{Binomial}\!\left(n,\, p\right) = \text{Binomial}\!\left(n,\, \frac{\lambda t}{n}\right)

Here nn is the number of subintervals (trials) and p=λtnp = \frac{\lambda t}{n} is the probability of exactly one event in each subinterval. Notice that as nn grows, pp shrinks proportionally. We continue to slice the interval ever finer while keeping the total expected count np=λtnp = \lambda t fixed.

The probability of exactly kk events is:

P(X=k)=(nk)(p)k(1p)nk,where p=λtnP(X = k) = \binom{n}{k} \left(p\right)^k \left(1 - p\right)^{n-k}, \quad \text{where} \space p = \frac{\lambda t}{n}

P(X=k)=(nk)(λtn)k(1λtn)nkP(X = k) = \binom{n}{k} \left(\frac{\lambda t}{n}\right)^k \left(1 - \frac{\lambda t}{n}\right)^{n-k}

Step 4: Take the limit. Let nn \to \infty — infinitely many, infinitely thin subintervals. Each factor converges separately:

(nk)1nk    1k!(λtn)knk=(λt)k(1λtn)n    eλt\binom{n}{k} \cdot \frac{1}{n^k} \;\to\; \frac{1}{k!} \qquad \left(\frac{\lambda t}{n}\right)^k \cdot n^k = (\lambda t)^k \qquad \left(1 - \frac{\lambda t}{n}\right)^n \;\to\; e^{-\lambda t}

Multiplying the pieces together:

P(X=k)=(λt)keλtk!,kN0P(X = k) = \frac{(\lambda t)^k \, e^{-\lambda t}}{k!}, \quad k \in \mathbb{N}_0

This is the Poisson distribution.

The Poisson Distribution

The Poisson probability mass function (PMF) has one parameter: μ=λt\mu = \lambda t, the expected number of events in the interval.

Note: We say PMF rather than PDF because kk is discrete. You can have 0, 1, 2, or 24 customers, but you can’t have 3.7 customers. The exponential distribution used a PDF because time is continuous and can take any non-negative real value.

P(X=k)=μkeμk!P(X = k) = \frac{\mu^k \, e^{-\mu}}{k!}

Its key properties:

PropertyFormula
Mean (expected value)E[X]=λtE[X] = \lambda t
VarianceVar(X)=λt\text{Var}(X) = \lambda t
Standard deviationσ=λt\sigma = \sqrt{\lambda t}

The mean and variance are equal. This is the fingerprint of the Poisson distribution — if you fit a Poisson model to real data and find the sample mean and variance are very different, that’s a signal the Poisson isn’t the right model.

Back to the coffee shop. With λ=3\lambda = 3 and t=8t = 8, we get μ=λt=24\mu = \lambda t = 24. Over a full shift, the expected number of customers is 24. But how does the probability spread?

P(X=20)=2420e2420!0.0624P(X = 20) = \dfrac{24^{20} \, e^{-24}}{20!} \approx 0.0624

P(X=24)=2424e2424!0.0812P(X = 24) = \dfrac{24^{24} \, e^{-24}}{24!} \approx 0.0812

P(X=30)=2430e2430!0.0363P(X = 30) = \dfrac{24^{30} \, e^{-24}}{30!} \approx 0.0363

Even the most likely outcome (k=24k = 24) only occurs about 8% of the time. The probability spreads across a wide range. That’s typical for a Poisson distribution with a large mean — it looks roughly bell-shaped, which is no accident (the central limit theorem explains why).

Interactive PMF

Use the slider to see how μ=λt\mu = \lambda t shapes the distribution.

mean = 5, variance = 5

A few things to notice:

Two Sides of the Same Coin

The exponential distribution and the Poisson distribution describe the same underlying process from two different perspectives. Both stem from the same assumption: events arrive independently at a constant average rate λ\lambda.

QuestionDistributionType
How long until the next event?Exp(λ)\text{Exp}(\lambda)Continuous
How many events in time tt?Poisson(λt)\text{Poisson}(\lambda t)Discrete

If you know λ\lambda for one, you know it for the other. The two distributions are mathematically inseparable. They are two ways of viewing the same process.

Real-World Examples

Anywhere events arrive at a roughly constant, independent rate, you can count them with the Poisson distribution:

The same caveats apply as with the exponential: the Poisson is the right model when events are independent and the rate is roughly constant. If customers tend to arrive in groups, or if the rush-hour rate is three times the off-peak rate, a more nuanced model is needed.