Stochastic OR

1.2 Exponential and Poisson Distribution

In this section, we make a model for the arrival process that is the most useful for queueing and inventory systems.1 In queueing systems, an arrival is often called customer or job, while in inventory systems it is called demand. In particular, we show how the exponential and gamma distributions relate to the Poisson distribution, and we use the memoryless property of job inter-arrival times as our point of departure.

We start with two examples to illustrate why it is reasonable to model the time between two events as memoryless. For instance, assume an hour ago a patient with a broken arm arrived at the first aid of a hospital. What can we say about the time the next patient with a broken arm will arrive? In other words, does the information that an hour ago a broken arm arrived help to make a better prediction of the next? Similarly, suppose that the average time between the sales of some shirt in a clothing store is 5 minutes. The shop owner informs us that no shirt has been sold in the past 7 minutes. Surely, this knowledge does not imply that the next shirt will necessarily be sold within the next minute.

Let us formalize this. If a person does not tell us anything about the past, we write \(\P{X>t}\) for the probability that no event occurs before time \(t\). However, if the person tells us at time \(s\) there was no arrival during \([0,s]\), we write \(\P{X>s+t\given X>s}\) for the probability that there will not be an arrival for an additional amount of time \(t\). If we believe that the information \(\{X>s\}\) is immaterial to predicting the lifetime of \(X\) after \(s\), then necessarily \(\P{X > s+t \given X>s} = \P{X>t}\).

Def. 1.2.1

A rv2 rv(s) := random variable(s). \(X\) whose cdf3 cdf := cumulative distribution function. satisfies

\begin{equation*} \P{X > s+t \given X>s} = \P{X>t}, \quad s, t \geq 0, \tag{1.2.1} \end{equation*}

is said to be memoryless.4 Not all rvs related to lifetime are memoryless. As a counterexample, take \(X\) as the lifespan of an arbitrary person born in 1820, and \(s= 100\) and \(t = 99\) years. Then \(\P{X>99}\) was small, but not zero, but \(\P{X>199 \given X>100} = 0\).

The next theorem is fundamental.

Th. 1.2.2

\(X\sim \Exp{\lambda}\) iff \(X\) is memoryless.

Proof

Observing that \(\P{X \leq t} = 1- e^{-\lambda t} \implies \P{X>t} = e^{-\lambda t}\),

\begin{align*} \P{X > s+t \given X>s} &\stackrel1= \frac{\P{X>s+t}}{\P{X>s}} = \frac{e^{-\lambda(t+s)}}{e^{-\lambda s}} \\ &= e^{-\lambda t} =\P{X>t}, \end{align*}

where Step 1 follows from the standard formula for conditional probability and that \(\{X>t+s\} \subset \{X>s\}\). The other direction is somewhat more difficult; we skip it. 5 For a proof, see A. A. Yushkevich and E. B. Dynkin (1969), Markov Processes: Theorems and Problems, Appendix 3.

Note that for \(X\sim \Exp{\lambda}\), the following properties hold:6 See Ex 1.2.6 for details.

\begin{align*} \E X &= \lambda^{-1}, & \V X &= \lambda^{-2}. \end{align*}

If we have a sequence of inter-arrival times \(\{X_{i}\}_{i=1}^{\infty}\), then we construct arrival times according to the recursive rule

\begin{align*} A_k &= A_{k-1} + X_k, & A_{0} = 0. \tag{1.2.2} \end{align*}

When inter-arrival times \(\{X_{k}\}\) are iid7 iid := independent and identically distributed. rvs and distributed as the common rv8 Henceforth, we write \(X_{k}\) iid \(\sim X\), or \(X_{k}\) iid \(\sim \Exp{\lambda}\). \(X\sim \Exp{\lambda}\), it is clear that the density of the first arrival time is

\begin{equation*} f_{A_{1}}(t) = f_{X} (t) = \lambda e^{-\lambda t} = \lambda \frac{(\lambda t)^{0}}{0!} e^{-\lambda t}, \end{equation*}

where we add the last term to see how a pattern emerges. Using convolution, for the second arrival time,

\begin{align*} f_{A_{2}}(t) &=\int_{0}^{t} f_{A_1}(s) f_{X}(t-s)\d s = \int_{0}^{t} \lambda e^{-\lambda s} \lambda e^{-\lambda(t-s)}\d s = \lambda \frac{(\lambda t)^1}{1!} e^{-\lambda t}. \end{align*}

With induction it is straightforward to extend the pattern and prove that \(A_{k} \sim \Gamma{\lambda, k}\), i.e.,

\begin{equation*} f_{A_k}(t) = \int_{0}^{t}f_{A_{k-1}}(s) f_X(t-s)\d s = \lambda \frac{(\lambda t)^{k-1}}{(k-1)!} e^{-\lambda t}. \tag{1.2.3} \end{equation*}

When the arrival times 9 Arrival times are not the same as inter-arrival times. are gamma distributed, the cdf of the number \(N(t)\) of jobs arriving during \([0, t]\) follows readily by observing that

\begin{equation*} \P{N(t) = k} = \P{A_{k} \leq t < A_{k+1}} = \P{A_k \leq t} - \P{A_{k+1} \leq t}, \tag{1.2.4} \end{equation*}

that is, to have precisely \(k\) arrivals during \([0, t]\), the arrival time \(A_{k}\) of job \(k\) must lie in \([0, t]\), but job \(k+1\) must arrive after \(t\). Using the cdf of \(A_{k}\) and integration by parts,

\begin{align*} \P{A_{k+1} \leq t} &= \lambda \int_0^t \frac{(\lambda s)^{k}}{k!}e^{-\lambda s}\, \d s = \lambda \frac{(\lambda s)^{k}}{k!}\frac{e^{-\lambda s}}{-\lambda} \Big|_{0}^t + \lambda \int_0^t \frac{(\lambda s)^{k-1}}{(k-1)!}e^{-\lambda s}\, \d s \\ &= - \frac{(\lambda t)^{k}}{k!} e^{-\lambda t} + \P{A_k \leq t}. \end{align*}

Combining this with Eq. (1.2.4), we find

\begin{equation*} \P{N(t) = k} = \frac{(\lambda t)^{k}}{k!} e^{-\lambda t}, \tag{1.2.5} \end{equation*}

that is, \(N(t) \sim \Pois{\lambda t}\), i.e., Poisson distributed with parameter \(\lambda t\). It is simple to see from Eq. (1.2.5) that10 See Ex 1.2.8 for details.

\begin{align*} \E{N(t)} &= \lambda t & \V{N(t)} &= \lambda t. \end{align*}

Interestingly, we can use Eq. (1.2.4) and Eq. (1.2.5) to derive the pdf of \(A_{k}\) from the Poisson distribution. From Eq. (1.2.5), \(\P{N(t) = 0} = e^{-\lambda t}\) when \(k=0\), and as \(A_0=0\) by definition, the RHS11 RHS := right hand side, LHS := left hand side. of Eq. (1.2.4) becomes \(1-\P{A_1\leq t}\). Consequently, \(\P{A_{1}\leq t} = 1 - e^{-\lambda t}\), and by differentiating this, the pdf is seen to be \(f_{A_1}(t) = \lambda e^{-\lambda t}\). For \(k\geq 1\), take the derivative with respect to \(t\) of Eq. (1.2.4) to see that \(\d \P{N(t)=k}/\d t = f_{A_{k}}(t) - f_{A_{k+1}}(t)\). It then follows from Eq. (1.2.5) that

\begin{equation*} \lambda \frac{(\lambda t)^{k-1}}{(k-1)!} e^{-\lambda t} - \lambda \frac{(\lambda t)^{k}}{k!} e^{-\lambda t} = f_{A_{k}}(t) - f_{A_{k+1}}(t). \end{equation*}

With recursion, we retrieve Eq. (1.2.3).

The family of rvs \(N_\lambda=\{N(t), t\geq 0\}\) is a Poisson process12 A random process is a much more complicated mathematical object than a rv: a process is a (possibly uncountable) set of rvs indexed by time, not just one rv. with arrival rate \(\lambda\). Once we have \(N_{\lambda}\), we can define \(N(s, t] := N(t)-N(s)\) as the number of customers that arrive in a general time period \((s, t]\).13 Note that \([0,t]\) is closed at both ends, but \((s,t]\) is open at the left.

It is important to know that \(N_\lambda\) has stationary and independent increments. Stationarity means that the distributions of the number of arrivals are the same for all intervals of equal length, that is, \(N(s,t]\) has the same probability distribution as \(N(u, v]\) if \(t-s = v-u\). Independence means, roughly speaking, that knowing that \(N(s,t]= n\) does not help to make any predictions about the value of \(N(u, v]\) if the intervals \((s,t]\) and \((u, v]\) do not overlap.

Eq. (1.2.5) implies a number of useful properties of the Poisson process. With small \(o\) notation14 A function \(f(h) = o(h)\) when \(f(h)/h \to 0\) as \(h\to 0\), e.g., \(x^{2} = o(x)\). , if \(t\ll 1\),

\begin{equation*} \begin{aligned} \P{N( t) = 0} &= e^{-\lambda t} = 1-\lambda t + o( t), \\ \P{N( t) = 1} &= \lambda t e^{-\lambda t} = \lambda t (1-\lambda t + o(t)) = \lambda t + o( t), \\ \P{N( t) \geq 2} &= 1 - \P{N(t) = 0} - \P{N(t)= 1} \\ &= 1 - (1-\lambda t) - \lambda t + o(t) = o(t). \end{aligned} \tag{1.2.6} \end{equation*}

It is important to know that the reverse of the above holds, that is, a stochastic process with stationary and independent increments and satisfying Eq. (1.2.6) is necessarily a Poisson process.15 We refer to the literature on (mathematical) probability theory for further background.

When we merge two independent Poisson processes \(N_{\lambda}\) and \(N_{\mu}\), e.g., adults and children entering a store16 Or some other feature by which you want to distinguish between customers. , we obtain a new Poisson process \(N_{\lambda +\mu}\) with rate \(\lambda + \mu\). To see this, we use the above. For \(t\ll 1\),

\begin{equation*} \begin{aligned} \P{N_{\lambda+\mu}( t) = 0} &= \P{N_{\lambda}( t) = 0} \P{N_{\mu}( t) = 0} \\ &= (1-\lambda t)(1-\mu t) + o(t) \\ &= 1 - (\lambda + \mu)t + o(t), \\ \P{N_{\lambda+\mu}( t) = 1} &= \P{N_{\lambda}( t) = 1} \P{N_{\mu}( t) = 0} \\ &\quad + \P{N_{\lambda}( t) = 0} \P{N_{\mu}( t) = 1} \\ &= \lambda t (1-\mu t) + (1-\lambda t)\mu t + o(t)\\ &= (\lambda + \mu) t + o(t), \end{aligned} \tag{1.2.7} \end{equation*}

and the rest of Eq. (1.2.6) follows similarly. So we see that we obtain the same type of equations as in Eq. (1.2.6) but with \(\lambda\) replaced by \(\lambda+\mu\). Realizing that \(N_{\lambda}\) and \(N_{\mu}\) have stationary and independent increments, it follows that \(N_{\lambda +\mu}\) is a Poisson process with rate \(\lambda+\mu\).

We can also split, or thin, a Poisson process into several substreams. For instance, consider a Poisson process \(N_\lambda\) of people passing by a shop, and suppose that some arbitrary customer decides to enter the shop as a Bernoulli distributed rv \(B\), independent of \(N_{\lambda}\), with success probability \(p\) and failure probability \(q=1-p\). Then the process of customers entering the shop is a Poisson process \(N_{\lambda p}\) with rate \(\lambda p\), because, for \(t\ll 1\), from Eq. (1.2.6),17 Why are there \(o(t)\) symbols already after the first equality signs? (Hint, it can happen that \(N(t)=2\), but both persons do not enter.)

\begin{align*} \P{N_{\lambda p}(t) = 0 } &= \P{N_{\lambda}(t) = 0} +\P{N_{\lambda}(t) = 1}\P{B=0} + o(t)\\ &= 1- \lambda t + \lambda t q + o(t) = 1 - \lambda p t + o(t),\\ \P{N_{\lambda p}(t) = 1 } &= \P{N_{\lambda}(t) = 1}\P{B=1} + o(t)= \lambda t p + o(t),\\ \P{N_{\lambda p}(t) \geq 2 } &\leq \P{N_{\lambda}(t) \geq 2} = o(t) \end{align*}

The resulting process \(N_{\lambda p}\) has an arrival rate of \(\lambda p\) and has stationary and independent increments because \(N_{\lambda}\) is a Poisson process.

The concepts of merging and thinning are useful to analyze queueing networks. Suppose the departure stream of a machine splits into two substreams, e.g., a fraction \(p\) of the jobs moves on to another machine and the rest (\(1-p\)) of the jobs leaves the system. Then we can model the arrival stream at the second machine as a thinned stream (with probability \(p\)) of the departures of the first machine. Merging occurs where the output streams of various stations are sent to a common downstream station.

A final interesting property of the Poisson process is the distribution of arrivals over an interval \([0,t]\) given that \(N(t) = n\).

Th. 1.2.3

Conditional on the event \(N(t)=n\), \(t>0\), the distribution of the arrival times \(\{A_{i}\}_{i=1}^{n}\) is the same as the distribution of the order statistic of \(n\) iid rvs \(\sim \Unif{[0,t]}\).18 \(\Unif{A}\) is the uniform distribution on the set \(A\).

Proof

We write for convenience \(\d t_i = [t_{i}, t_i + \d t]\), and similarly \(\d(t_i-t_{i-1}) = [t_i-t_{i-1}, t_i-t_{i-1}+\d t]\).19 This notation allows us to write \(\P{X_i \in \d t_i} = \lambda e^{-\lambda t_i} \d t\) when \(X_i\sim \Exp{\lambda}\), and \(\P{U_i\in \d t_i } = \d t / t\) when \(U_i\sim \Unif{[0,t]}\). We provide the proof for \(N(t)=2\); the proof for the general case is nearly identical, but needs more notation.

Consider the Poisson process at \(0\leq t_1<t_2 \leq t\). Then,

\begin{equation*} \P{A_1\in \d t_1, A_2\in \d t_2 \given N(t)=2} = \frac{\P{A_1\in \d t_1, A_2\in \d t_2, A_3 > t}}{\P{N(t)=2}}, \tag{\(\star\)} \end{equation*}

because \(N(t) = 2 \iff A_{2} \leq t < A_{3}\). For the numerator,

\begin{align*} \P{A_1\in \d t_1, A_2\in \d t_2, A_3 > t} &\stackrel1= \P{X_1\in \d t_1, X_2\in \d (t_2-t_1), X_3 > t-t_{2}} \\ &\stackrel2= \lambda e^{-\lambda t_1} \d t\lambda e^{-\lambda (t_2-t_1)} \d te^{-\lambda (t-t_2)} \\ &\stackrel3= \lambda^{2} e^{-\lambda t} \d t^{2}, \end{align*}

where 1 follows from the recursive relation \(A_{i} = A_{i-1} + X_{i}\); 2 because the \(X_i\) are iid \(\sim \Exp{\lambda}\); 3 is algebra. Because \(\P{N(t)=2} = (\lambda t)^2 e^{-\lambda t}/2!\), we find that \(\P{A_1\in \d t_1, A_2\in \d t_2 \given N(t)=2} = 2! \d t^{2}/t^{2}\).

As for the iid uniform rvs, \(\P{U_1\in \d t_1, U_2\in \d t_2} = \d t^2/ t^{2}\). However, there are \(2!\) ways that the uniform rvs can hit the set \(\d t_1 \d t_{2}\), so the probability that the order statistic of \(U_{1}, U_2\) lies in this set is \(2! \d t^{2}/t^{2}\). This equals the conditional probability in \((\star)\).

Summarizing, the Poisson process \(N_{\lambda}\) and the exponential distribution are very closely related: a counting process \(\{N_{\lambda}(t)\}\) is a Poisson process with rate \(\lambda\) if and only if the inter-arrival times \(\{X_k\}\) are iid and \(X_{k} \sim \Exp{\lambda}\). Thus, if you find it reasonable to model inter-arrival times as memoryless, then the number of arrivals in an interval is necessarily Poisson distributed. Moreover, if you find it reasonable that the occurrence of an event in a small time interval is equally likely over time and independent from one interval to another, then the arrival process is Poisson, and the inter-arrival times are exponential.

TF 1.2.1

Let \(N\) be a Poisson process with rate \(\lambda\) and \(0<s<t\). Claim:

\begin{equation*} \{N(0,s]+N(s,t]=1\}\cap\{N(0,t]=1\} = \{N(0,s]=1\}. \end{equation*}
Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

False. Since \(N(0,s] + N(s, t] = N(0,t]\), the LHS just says that there was one arrival during \((0,t]\). The RHS says something different, in particular because \(s<t\).

TF 1.2.2

Assume that \(N_a(t)\sim \Pois{\lambda t}\) , \(N_s(t) \sim \Pois{\mu t}\) and independent. Claim:

\begin{equation*} \P{N_a(t) + N_s(t) = n} = e^{-(\mu+\lambda)t} \sum_{i=0}^n \frac{(\mu t)^{n-i}}{(n-i)!} \frac{(\lambda t)^i}{i!}. \\ \end{equation*}
Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

True.

TF 1.2.3

Claim: this program prints 6.

Python
X = [0, 1, 2, 3, 4, 7]
A = [1] * len(X)

for i in range(2, 5):
    A[i] += A[i - 1] + X[i]

print(A[3])
Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

False, it prints 8. To see why, observe that for \(i=2\), \(A[i-1] = A[1] = 1\) and \(X[2] = 2\). Thus, for \(i=2\) the RHS in the loop becomes \(3\). Now there is the \(+=\) symbol, which implies that \(3\) is added to \(A[2]\). As the latter is 1, \(A[2]\) becomes \(1+ 3 = 4\) after the first iteration of the for loop. In the second iteration, we, therefore, find that \(4+3\) gets added to \(A[3]=1\), so \(A[3]\) becomes 8.

TF 1.2.4

The inter-arrival times \(\{X_k\}\) are iid and exponentially distributed with mean \(1/\lambda\), and \(A_k = \sum_{i=1}^k X_i\), \(A_{0} = 0\). Claim:

\begin{equation*} \P{A_{k+1} \leq t} = - \frac{(\lambda t)^{k}}{k!} e^{-\lambda t} + \P{A_k \leq t}. \end{equation*}
Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

True. Observe that

\begin{equation*} \P{A_k \leq t} - \P{A_{k+1} \leq t} = \frac{(\lambda t)^{k}}{k!} e^{-\lambda t}. \end{equation*}
TF 1.2.5

Claim: this program prints 31.

Python
tot, thres = 8, 30
while tot < thres:
    tot += 5

print(tot)
Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

False, it prints 33, because the algorithm keeps adding \(5\) to 8 until the threshold 30 is exceeded. It does not print the intermediate values of tot because print(tot) is not part of the body of the loop.

The exercises require pen and paper. All problems have solutions, but not all have hints.

Ex 1.2.6

Show that when \(X\sim \Exp{\lambda}\) its moment generating function (mgf) is \(M_X(t) := \E{e^{tX}} = \lambda/(\lambda -t)\), and use this to calculate \(\E X\) and \(\V X\).

Hint

Recall that \(\E X = M_{X}'(0)\), \(\E{X^{2}} = M_{X}''(0)\), and \(\V X = \E{X^{2}} - (\E X)^{2}\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

For \(t< \lambda\), and \(f(t) = \lambda e^{-\lambda t} \1{t\geq 0}\),

\begin{align*} M_X(t) &= \E{\exp(t X)} =\int_0^\infty e^{tx} f(x) \,\d x =\int_0^\infty e^{tx} \lambda e^{-\lambda x} \,\d x =\frac{\lambda}{\lambda -t}. \end{align*}

This last integral only converges when \(\lambda > t\). Next, \(M_X'(t)=\lambda/(\lambda-t)^2 \implies M_X'(0)=1/\lambda\), \(M_X''(t)=2\lambda/(\lambda-t)^3 \implies \E{X^2}=2\lambda^{-2}\). Thus, \(\V X = \E{X^2} - (\E X)^2 = \frac{2}{\lambda^2} - (\frac{1}{\lambda})^2 = \lambda^{-2}\).

Ex 1.2.7

Use mgfs to prove that \(A_i\) has density \(f_{A_i}(t) = \lambda e^{-\lambda t} \frac{(\lambda t)^{i-1}}{(i-1)!}\).

Hint

Use that for independent rvs \(X, Y\), \(M_{X+Y}(t)=M_{X}(t) M_{Y}(t)\). Why is \(M_{A_i}(t) = \E{e^{t A_i}} = \prod_{k=1}^{i} \E{e^{tX_k}}\)?

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

Using the iid property of the \(\{X_i\}\),

\begin{align*} M_{A_i}(t) &= \E{e^{t A_i}} = \E{\exp\left(t\sum_{k=1}^{i} X_k\right)} = \prod_{k=1}^{i} \E{e^{tX_k}} = \left(\frac{\lambda}{\lambda -t }\right)^i. \end{align*}

From a table of moment-generating functions, it follows immediately that \(A_i \sim \Gamma{(i,\lambda)}\), i.e., \(A_i\) is Gamma distributed.

Ex 1.2.8

When \(N \sim \Pois{\lambda}\) and \(\alpha > 0\), show that \(\E{\alpha^{N}} = e^{\lambda(\alpha-1)}\). Use this to see that \(M_{N(t)}(s) = \exp{(\lambda t(e^s-1))}\). Then find \(\E{N(t)}\) and \(\V{N(t)}\).

Hint

LOTUS: \(\E{\alpha^{N}} = \sum_{k=0}^\infty \alpha^{k} \P{N=k}\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real
\begin{align*} \sum_{k=0}^\infty \alpha^{k} \P{N=k} = \sum_{k=0}^\infty \alpha^{k} \frac{(\lambda)^k}{k!} e^{-\lambda} = e^{-\lambda} \sum_{k=0}^\infty \frac{(\alpha \lambda)^k}{k!} =\exp(\lambda (\alpha - 1)). \end{align*}

Taking \(\alpha=e^{s}\) in this equation and differentiating gives

\begin{equation*} M_{N(t)}'(s) = \lambda t e^s \exp(\lambda t(e^s - 1)). \end{equation*}

Hence \(\E{N(t)} = M_{N(t)}'(0) = \lambda t\). Next, \(M_{N(t)}''(s) = (\lambda t e^s + (\lambda t e^s)^2) \exp(\lambda t(e^s - 1))\), hence \(\E{(N(t))^2} = M_{N(t)}''(0) = \lambda t + (\lambda t)^2\), and thus, \(\V{N(t)} =\E{(N(t))^2}-(\E{N(t)})^2 = \lambda t + (\lambda t)^2 - (\lambda t)^2 = \lambda t\).

Ex 1.2.9

If \(X\sim\Exp{\lambda}, S\sim\Exp{\mu}\) and are independent, show that \(Z=\min\{X,S\}\sim\Exp{\lambda+\mu}\), hence \(\E Z = (\lambda+\mu)^{-1}\).

Hint

This result can be anticipated when you think about merging Poisson processes. Then, use \(\{\min\{X, S\}>x\} =\{X>x\} \cap\{S> x\}\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

\(\P{Z>x} = \P{\min\{X,S\}>x} = \P{X>x, S> x} = \P{X>x}\P{S>x} = e^{-\lambda x} e^{-\mu x}\), since \(X\) and \(S\) are independent.

Ex 1.2.10

If the Poisson arrival processes \(N_\lambda\) and \(N_\mu\) are independent, show that

\begin{equation*} \P{N_\lambda(t) = 1 \given N_\lambda(t) + N_\mu(t) = 1} = \frac{\lambda}{\lambda+\mu}. \end{equation*}

In other words, given that a customer arrived in \([0,t]\), the probability that the customer is of the first type is \(\lambda/(\lambda+\mu)\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

From Eq. (1.2.5) and Eq. (1.2.7) and

\begin{align*} & \P{N_\lambda(t) = 1 \given N_\lambda(t) + N_\mu(t) = 1} = \frac{\P{N_\lambda(t) = 1}\P{N_\mu(t) = 0}}{\P{N_{\lambda+\mu}(t) = 1}} \\ &= \frac{\lambda t \exp(-\lambda t) \exp(-\mu t)}{((\lambda+\mu)t)\exp{(-(\lambda+\mu)t)}} = \frac{\lambda t \exp{(-(\lambda + \mu)t)}}{((\lambda+\mu)t)\exp{(-(\lambda+\mu)t)}} = \frac{\lambda}{\lambda+\mu}. \end{align*}
Ex 1.2.11

If \(X\sim \Exp{\lambda}\), \(S\sim\Exp{\mu}\) and are independent, show that

\begin{equation*} \P{X\leq S} = \frac{\lambda}{\lambda+\mu}. \end{equation*}
Hint

Think about splitting Poisson processes, and use that \(\P{S > X} = \E{\E{\1{S>X}|X}}\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real

With conditioning, \(\P{S>X \given X=t} = \P{S>t} =e^{-\mu t}\). With the fundamental bridge and conditional expectation, \(\E{\1{S>X} \given X} = e^{-\mu X}\), hence \(\E{\1{S>X}} = \E{e^{-\mu X}}\). But this last formula is equal to \(M_{X}(-\mu)\), so taking \(t=-\mu\) in the mgf of \(X\) we obtain \(\lambda/(\lambda+\mu)\).

If the somewhat heuristic derivations above do not satisfy your sense for rigor, then you should solve the next set of exercises; otherwise you can skip them.

Opt 1.2.12

If the Poisson arrival processes \(N_\lambda\) and \(N_\mu\) are independent, use moment generating functions to show that \(N_\lambda + N_\mu\) is a Poisson process with rate \(\lambda + \mu\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real
\begin{align*} M_{N_\lambda(t)+N_\mu(t)}(s) &= M_{N_\lambda(t)}(s)\cdot M_{N_{\mu}(t)}(s) =\exp(\lambda t (e^s -1))\cdot \exp(\mu t(e^s-1)) \\ &= \exp((\lambda + \mu)t (e^s-1)). \end{align*}
Opt 1.2.13

Show with moment-generating functions that thinning the Poisson process \(N_\lambda\) by means of Bernoulli rvs with success probability \(p\) results in a Poisson process \(N_{\lambda p}\).

Hint

Dropping the dependence of \(N\) on \(t\) for the moment for notational convenience, consider the rv

\begin{equation*} Y = \sum_{i=1}^N Z_i, \end{equation*}

with \(N\sim P(\lambda)\) and \(Z_i\sim B(p)\) and \(\{Z_{i}\}\) iid. Show that the moment-generating function of \(Y\) is equal to the moment-generating function of a Poisson rv with parameter \(\lambda p\).

Solution
Did you actually try? Maybe see the 'hints' above!:
Solution, for real
\begin{equation*} \E{e^{s Z}} = e^0 \P{Z=0} + e^{s} \P{Z=1} = (1-p) + e^s p, \end{equation*}

from which

\begin{equation*} \E{e^{s\sum_{i=1}^n Z_i}} = \left(\E{e^{s Z}}\right)^n = \left(1 + p (e^s - 1)\right)^n, \end{equation*}

where we use that the \(Z_i\) are iid. Thus, (see the hint) \(\E{e^{sY}\big|N=n} = \E{e^{s\sum_{i=1}^n Z_i}} = \alpha^{n}\) where \(\alpha =1 + p (e^s - 1)\). With Adam's law and Ex 1.2.8, we get

\begin{equation*} \E{e^{sY}} = \E{\alpha^{N}} = e^{\lambda (\alpha - 1)} = \exp(\lambda p (e^s - 1)). \end{equation*}

This is the mgf of a Poisson rv with rate \(\lambda p\).