Random walk

Basic defitions

One of the simplest random processes is so-called “simple random walk”.

Definition. Let {Xk}k=1 be a sequence of i.i.d. discrete random variables with finite outcomes. For each positive integer n, we let Sn denote the sum X1+X2++Xn. The sequence {Sn}n=1 is called a **random walk**.

Remark. If the common range of the Xk’s is Rm, then we say that {Sn} is a random walk in Rm.

Remark. Sometimes we set S0=a for some number a as the **start value** of the random walk. This is common in the gambling analysis.

Lemma. The simple random walk has the Markov property ; that is P(Sm+n=j|S0,S1,,Sm)=P(Sm+n=j|Sm),n0.

Proof ▸

If one knows the value of Sm , then the distribution of Sm+n depends only on the jumps Xm+1,,Xm+n, and cannot depend on further information concerning the values of S0,S1,,Sm1. ◼

The random walk has two main applications: gambling analysis and investment decision making. For example, in the gambling, we gamble at each step, Si is the total stack we have at step i, and we are interested in when there will be a return or \text{equalization}, i.e., what is the probability that S0=a=Sn.

Example. [Random walks on the real line] We shall first consider the simplest non-trivial case of a random walk in R1, namely the case where the common distribution function of the random variables Xn is given by (9-1)fX(x)={1/2, if x=±10, otherwise  We note that in this situation, all paths of length n have the same probability, namely 2n. We can plot the random walk in the plane, where the horizontal axis represents time and the vertical axis represents the value of Sn. Given a sequence {Sn} of partial sums, we first plot the points (n,Sn), and then for each k<n, we connect (k,Sk) and (k+1,Sk+1) with a straight line segment. The **length** of a path is just the difference in the time values of the beginning and ending points on the path. This can be visualized in Figure ???.

Fig.1: A random walk Sn.

Returns and First Returns

This subsection studies the random walk in R1. The p.m.f. of Xi is given by (9-1).

We say that an equalization has occurred, or there is a return to the origin at time n, if Sn=0. Note that the returns are possible only after even number of steps. To calculate the probability of an equalization at time 2m, we need only count the number of paths of length 2m which begin and end at the origin. The number of such paths is clearly (2mm)=(2m)!m!m!. Since each path has probability 22m, we have the following theorem.

Theorem. The probability of a return to the origin at time 2m is given by u2m=(2mm)22m. The probability of a return to the origin at an odd time is 0.

A random walk is said to have a first return to the origin at time 2m if m>0, and S2k=0 for all k<m. We define f2m to be the probability of this event. (We also define f0=0.) One can think of the expression f2m22m as the number of paths of length 2m between the points (0,0) and (2m,0) that do not touch the horizontal axis except at the endpoints. We have the following theorem.

Theorem. For n1, the probabilities {u2k} and {f2k} are related by the equation u2n=f0u2n+f2u2n2++f2nu0, where f0=0 and u0=1 as convention.

Proof ▸

There are u2n22n paths of length 2n which have endpoints (0,0) and (2n,0). The collection of such paths can be partitioned into n sets, depending upon the time of the first return to the origin. A path in this collection which has a first return to the origin at time 2k consists of an initial segment from (0,0) to (2k,0), in which no interior points are on the horizontal axis, and a terminal segment from (2k,0) to (2n,0), with no further restrictions on this segment. Thus, the number of paths in the collection which have a first return to the origin at time 2k is given by f2k22ku2n2k22n2k=f2ku2n2k22n. If we sum over k, we obtain the equation u2n22n=f0u2n22n+f2u2n222n++f2nu022n Dividing both sides of this equation by 22n completes the proof. ◼

Using the generating functions, we have following theorem.

Theorem. For m1, the probability of a first return to the origin at time 2m is given by f2m=u2m2m1=(2mm)(2m1)22m.

Proof ▸

We begin by defining the generating functions U(x)=m=0u2mxm and F(x)=m=0f2mxm. Note that two sequences converge when |x|<1. The preceding theorem says that U(x)=1+U(x)F(x). (The presence of the 1 on the right-hand side is due to the fact that u0 is defined to be 1, but the last theorem only holds for m1.) We can solve the above equation for F(x), obtaining F(x)=U(x)1U(x). Now, if we can find a closed-form expression for the function U(x), we will also have a closed-form expression for F(x). From the first theorem in this notes, we have U(x)=m=0(2mm)22mxm=m=0(2mm)(x4)m. From series expansion, we find that 114x=m=0(2mm)xm. Therefore, we have U(x)=11x,F(x)=U(x)1U(x)=11x. Although it is possible to compute the value of f2m using the Binomial Theorem, it is easier to note that F(x)=U(x)/2, so that the coefficients f2m can be found by integrating the series for U(x). We obtain, for m1, f2m=u2m22m=(2m2m1)m22m1=(2mm)(2m1)22m=u2m2m1, since (2m2m1)=m2(2m1)(2mm). This completes the proof of the theorem. ◼

Probability of Eventual Return

In the symmetric random walk process in Rm, we first examine this question in the case that m=1, and then we consider the general case.

We should be careful when dealing with eventual return since the sample space seems to be the set of all walks of infinite length, and this set is non-denumerable. To avoid difficulties, we will define wn to be the probability that a first return has occurred no later than time n. Thus, wn concerns the sample space of all walks of length n, which is a finite set. Then it is reasonable to define w=limnwn. This limit clearly exists and is at most one, since the sequence wnn=1 is an increasing sequence, and all of its terms are at most 1. In terms of the fn probabilities, we see that w2n=i=1nf2i. Thus w=i=1f2i. In previous proof we introduced the the generating function F(x)=m=0f2mxm=11x. which is convergent for |x|<1. In fact, it also converges for x=±1. Hence we see that w=F(1)=1. Thus, with probability one, the particle returns to the origin in R1 random walk.

Now we consider the eventual return in Rm. We define f2n(m) to be the probability that the first return to the origin in Rm occurs at time 2n. The quantity u2n(m) is defined in a similar manner. From the preceding theorem, we have u2n(m)=f0(m)u2n(m)+f2(m)u2n2(m)++f2n(m)u0(m). We continue to generalize previous work by defining U(m)(x)=n=0u2n(m)xn,F(m)(x)=n=0f2n(m)xn. Then we see that U(m)(x)=1+U(m)(x)F(m)(x). These functions will always converge in the interval (1,1). In fact, since w(m)=n=0f2n(m)1 for all m, the series for F(m)(x) converges at x=1 as well, and F(m)(x) is left continuous at x=1, i.e., limx1F(m)(x)=F(m)(1). Thus, we have (9-2)w(m)=limx1F(m)(x)=limx1U(m)(x)1U(m)(x), so to determine w(m), , it suffices to determine limx1U(m)(x). We let u(m) denote this limit. Then we have (9-3)u(m)=n=0u2n(m).

In R2 random walk, we have u2n(2)=142n(2nn)2. Using Stirling’s Formula (Stirling)n!nnen2πnfor n large. we have (2nn)22nπnRightarrowu2n(2)1πn. From this it follows easily that n=0u2n(2) diverges, so w(2)=1, i.e., in R2, the probability of an eventual return is 1.

In R3, we have u2n(3)=122n(2nn)j,k(13nn!j!k!(njk)!)2, Note that n=0u2n(3) converges, so w(3) is strictly less than one. This means that in R3, the probability of an eventual return to the origin is strictly less than one.

Gambler’s Ruin

In this section, we remove the assumption that the random walk is symmetric. Instead, we assume that p and q are non-negative real numbers with p+q=1, and that the common distribution function of the jumps of the random walk is fX(x)={p, if x=1q, if x=1

We formulate the problem as follows. A gambler starts with a “stake” of size s. She plays until her capital reaches the value M or the value 0. In the language of Markov chains, these two values correspond to absorbing states. We are interested in studying the probability of occurrence of each of these two outcomes.

We begin by defining Sk to be the probability that the gambler’s stake reaches 0, i.e., she is ruined, before it reaches M, given that the initial stake is k. We note that S0=1 and SM=0. The fundamental relationship among the Sk’s is the following: Sk=pSk+1+qSk1, where 1kM1. This holds because if her stake equals k, and she plays one game, then her stake becomes k+1 with probability p and k1 with probability q. Since p+q=1, we can rewrite the equation as p(Sk+1Sk)=q(SkSk1) or Sk+1Sk=qp(SkSk1). From this equation, it is easy to see that Sk+1Sk=(qp)k(S1S0). We now use telescoping sums to obtain an equation in which the only unknown is S1. 1=SMS0=k=0M1(Sk+1Sk) =k=0M1(qp)k(S1S0)=(S1S0)k=0M1(qp)k. Then we can solve for Sj, 1jM: Sj=1(q/p)j1(q/p)M1.

  • Consider M>>j>>0, q/p>1, then Sj=1(q/p)jM1.
  • Consider q/p<1, then Sj=(q/p)j0.