1. Random
  2. 15. Markov Processes
  3. 1
  4. 2
  5. 3
  6. 4
  7. 5
  8. 6
  9. 7
  10. 8
  11. 9
  12. 10
  13. 11
  14. 12
  15. 13
  16. 14
  17. 15
  18. 16
  19. 17
  20. 18
  21. 19
  22. 20
  23. 21
  24. 22
  25. 23

11. Discrete-Time Branching Chains

Basic Theory

Introduction

Generically, suppose that we have a system of particles that can generate or split into other particles of the same type. Here are some typical examples:

We assume that each particle, at the end of its life, is replaced by a random number of new particles that we will refer to as children of the original particle. Our basic assumption is that the particles act independently, each with the same offspring distribution on N. Let f denote the common probability density function of the number of offspring of a particle. We will also let fn=fff denote the convolution power of degree n of f; this is the probability density function of the total number of children of n particles.

We will consider the evolution of the system in real time in our study of continuous-time branching chains. In this section, we will study the evolution of the system in generational time. Specifically, the particles that we start with are in generation 0, and recursively, the children of a particle in generation n are in generation n+1.

Generations 0, 1, 2, and 3 of a branching chain.
The first three generations of a branching chain

Let Xn denote the number of particles in generation nN. One way to construct the process mathematically is to start with an array of independent random variables (Un,i:nN,iN+), each with probability density function f. We interpret Un,i as the number of children of the ith particle in generation n (if this particle exists). Note that we have more random variables than we need, but this causes no harm, and we know that we can construct a probability space that supports such an array of random variables. We can now define our state variables recursively by Xn+1=i=1XnUn,i

X=(X0,X1,X2,) is a discrete-time Markov chain on N with transition probability matrix P given by P(x,y)=fx(y),(x,y)N2 The chain X is the branching chain with offspring distribution defined by f.

Details:

The Markov property and the form of the transition matrix follow directly from the construction of the state variables given above. Since the variables (Un,i:nN,iN+) are independent, each with PDF f, we have P(Xn+1=yX0=x0,,Xn1=xn1,Xn=x)=P(i=1xUn,i=y)=fx(y)

The branching chain is also known as the Galton-Watson process in honor of Francis Galton and Henry William Watson who studied such processes in the context of the survival of (aristocratic) family names. Note that the descendants of each initial particle form a branching chain, and these chains are independent. Thus, the branching chain starting with x particles is equivalent to x independent copies of the branching chain starting with 1 particle. This features turns out to be very important in the analysis of the chain. Note also that 0 is an absorbing state that corresponds to extinction. On the other hand, the population may grow to infinity, sometimes called explosion. Computing the probability of extinction is one of the fundamental problems in branching chains; we will essentially solve this problem in the next subsection.

Extinction and Explosion

The behavior of the branching chain in expected value) is easy to analyze. Let m denote the mean of the offspring distribution, so that m=x=0xf(x) Note that m[0,]. The parameter m will turn out to be of fundamental importance.

Expected value properties

  1. E(Xn+1)=mE(Xn) for nN
  2. E(Xn)=mnE(X0) for nN
  3. E(Xn)0 as n if m<1.
  4. E(Xn)=E(X0) for each nN if m=1.
  5. E(Xn) as n if m>1 and E(X0)>0.
Details:

For part (a) we use a conditioning argument and the construction above. For xN, E(Xn+1Xn=x)=E(i=1xUn,i|Xn=x)=E(i=1xUn,i)=mx That is, E(Xn+1Xn)=mXn so E(Xn+1)=E[E(Xn+1Xn)]=mE(Xn) Part (b) follows from (a) and then parts (c), (d), and (e) follow from (b).

Part (c) of [2] is extinction in the mean; part (d) is stability in the mean; and part (e) is explosion in the mean.

Recall that state 0 is absorbing (there are no particles), and hence {Xn=0 for some nN}={τ0<} is the extinction event (where as usual, τ0 is the time of the first return to 0). We are primarily concerned with the probability of extinction, as a function of the initial state. First, however, we will make some simple observations and eliminate some trivial cases.

Suppose that f(1)=1, so that each particle is replaced by a single new particle. Then

  1. Every state is absorbing.
  2. The equivalence classes are the singleton sets.
  3. With probability 1, Xn=X0 for every nN.
Details:

These properties are obvious since P(x,x)=1 for every xN.

Suppose that f(0)>0 so that with positive probability, a particle will die without offspring. Then

  1. Every state leads to 0.
  2. Every positive state is transient.
  3. With probability 1 either Xn=0 for some nN (extinction) or Xn as n (explosion).
Details:
  1. Note that P(x,0)=[f(0)]x>0 for xN, so every state leads to 0 in one step.
  2. This follows from (a). If xN+, then x leads to the absorbing state 0 with positive probability. Hence a return to x, starting in x, cannot have probability 1.
  3. This follows from (a) and (b). With probability 1, every positive state is visited only finitely many times. Hence the only possibilities are Xn=0 for some nN or Xn as n.

Suppose that f(0)=0 and f(1)<1, so that every particle is replaced by at least one particle, and with positive probability, more than one. Then

  1. Every positive state is transient.
  2. P(Xn as nX0=x)=1 for every xN+, so that explosion is certain, starting with at least one particle.
Details:
  1. Let xN+. Under the assumptions on f, state x leads to some state y>x but y does not lead back to x. Hence with positive probability, the chain starting in x will not return to x.
  2. This follows from (a) and that the fact that positive states do not lead to 0.

Suppose that f(0)>0 and f(0)+f(1)=1, so that with positive probability, a particle will die without offspring, and with probability 1, a particle is not replaced by more than one particle. Then

  1. Every state leads to 0.
  2. Every positive state is transient.
  3. With probability 1, Xn=0 for some nN, so extinction is certain.
Details:
  1. As before, P(x,0)=[f(0)]x>0 for xN, so x leads to 0 in one step.
  2. This follows from (a) and the fact that 0 is absorbing.
  3. Under the assumptions on f, state x leads to state y only if yx. So this follows from (a) and (b).

Thus, the interesting case is when f(0)>0 and f(0)+f(1)<1, so that with positive probability, a particle will die without offspring, and also with positive probability, the particle will be replaced by more than one new particles. We will assume these conditions for the remainder of our discussion. By [4], all states lead to 0 (extinction). We will denote the probability of extinction, starting with one particle, by q=P(τ0<X0=1)=P(Xn=0 for some nNX0=1)

The set of positive states N+ is a transient equivalence class, and the probability of extinction starting with xN particles is qx=P(τ0<X0=x)=P(Xn=0 for some nNX0=x)

Details:

Under the assumptions on f, from any positive state the chain can move 2 or more units to the right and one unit to the left in one step. It follows that every positive state leads to every other positive state. On the other hand, every positive state leads to 0, which is absorbing. Thus, N+ is a transient equivalence class.

Recall that the branching chain starting with xN+ particles acts like x independent branching chains starting with one particle. Thus, the extinction probability starting with x particles is qx.

The parameter q satisfies the equation q=x=0f(x)qx

Details:

This result follows from conditioning on the first state. q=P(τ0<X0=1)=x=0P(τ0<X0=1,X1=x)P(X1=xX0=1) But by the Markov property and the previous result, P(τ0<X0=1,X1=x)=P(τ0<X1=x)=qx and of course P(X1=xX0=1)=P(1,x)=f(x).

Thus the extinction probability q starting with 1 particle is a fixed point of the probability generating function Φ of the offspring distribution: Φ(t)=x=0f(x)tx,t[0,1] Moreover, from the general discussion of hitting probabilities in the section on recurrence and transience, q is the smallest such number in the interval (0,1]. If the probability generating function Φ can be computed in closed form, then q can sometimes be computed by solving the equation Φ(t)=t.

Φ satisfies the following properties:

  1. Φ(0)=f(0).
  2. Φ(1)=1.
  3. Φ(t)>0 for t(0,1) so Φ in increasing on (0,1).
  4. Φ(t)>0 for t(0,1) so Φ in concave upward on (0,1).
  5. m=limt1Φ(t).
Details:

These are basic properties of the probability generating function. Recall that the series that defines Φ is a power series about 0 with radius of convergence r1. A function defined by a power series is infinitely differentiable within the open interval of convergence, and the derivates can be computed term by term. So Φ(t)=x=1xf(x)tx1>0,t(0,1)Φ(t)=x=2x(x1)f(x)tx2>0,t(0,1) If r>1 then m=Φ(1). If r=1, the limit result is the best we can do.

Our main result is next, and relates the extinction probability q and the mean of the offspring distribution m.

The extinction probability q and the mean of the offspring distribution m are related as follows:

  1. If m1 then q=1, so extinction is certain.
  2. If m>1 then 0<q<1, so there is a positive probability of extinction and a positive probability of explosion.
Details:

These results follow from the properties of the PGF Φ in [9]. See the graphs below.

  1. If m1 the graph of Φ will cross the diagonal line on the interval [0,1] only at t=1.
  2. If m>1 the graph of Φ will cross the diagonal line on the interval [0,1] in two places: t=q(0,1) and t=1.
The case of certain extinction.
PGF1.png
The case of possible extinction and possible explosion.
PGF2.png

Computational Exercises

Consider the branching chain with offspring probability density function f given by f(0)=1p, f(2)=p, where p(0,1) is a parameter. Thus, each particle either dies or splits into two new particles. Find each of the following.

  1. The transition matrix P.
  2. The mean m of the offspring distribution.
  3. The generating function Φ of the offspring distribution.
  4. The extinction probability q.
Details:

Note that an offspring variable has the form 2I where I is an indicator variable with parameter p.

  1. For xN, fx is the PDF of 2U where U has the binomial distribution with parameters x and p. Hence P(x,y)=fx(y)=(xy/2)py/2(1p)xy/2,y{0,2,,2x}
  2. m=2p.
  3. Φ(t)=pt2+(1p) for tR.
  4. q=1 if 0<p12 and q=1pp if 12<p<1.
Graphs of tΦ(t) and tt when p=13
Graphs
Graphs of tΦ(t) and tt when p=23
Graphs

Consider the branching chain whose offspring distribution is the geometric distribution on N with parameter 1p, where p(0,1). Thus f(n)=(1p)pn for nN. Find each of the following:

  1. The transition matrix P.
  2. The mean m of the offspring distribution.
  3. The generating function Φ of the offspring distribution.
  4. The extinction probability q.
Details:
  1. For xN, fx is the PDF of the negative binomial distribution on N with parameter 1p. So P(x,y)=fx(y)=(x+y1x1)py(1p)x,yN
  2. m=p1p.
  3. Φ(t)=1p1pt for |t|<1p.
  4. q=1 if 0<p12 and q=1pp if 12<p<1.
Graphs of tΦ(t) and tt when p=13
Graphs
Graphs of tΦ(t) and tt when p=23
Graphs

Curiously, the extinction probability is the same as in [11].

Consider the branching chain whose offspring distribution is the Poisson distribution with parameter m(0,). Thus f(n)=emmn/n! for nN. Find each of the following:

  1. The transition matrix P.
  2. The mean m of the offspring distribution.
  3. The generating function Φ of the offspring distribution.
  4. The approximate extinction probability q when m=2 and when m=3.
Details:
  1. For xN, fx is the PDF of the Poisson distribution with parameter mx. So P(x,y)=fx(y)=emx(mx)yy!,yN
  2. The parameter m is the mean of the Poisson distribution, so the notation is consistent.
  3. Φ(t)=em(t1) for tR.
  4. q=1 if 0<m1. If m>1 then q is the solution in (0,1) of the equation em(q1)=q which can be expressed in terms of a special function known as the Lambert W function: q=1mW(mem) For m=2, q0.20319. For m=3, q0.059520.
Graphs of tΦ(t) and tt when m=12
Graphs
Graphs of tΦ(t) and tt when m=2
Graphs