One of the simplest random processes is so-called “simple random walk”.
Definition.
Let be a sequence of i.i.d. discrete random variables with finite outcomes. For each positive integer , we let denote the sum . The sequence is called a **random walk**.
Remark.
If the common range of the ’s is , then we say that is a random walk in .
Remark.
Sometimes we set for some number as the **start value** of the random walk. This is common in the gambling analysis.
Lemma.
The simple random walk has the Markov property ; that is
If one knows the value of , then the distribution of depends only on the jumps , and cannot depend on further information concerning the values of .
◼
The random walk has two main applications: gambling analysis and investment decision making. For example, in the gambling, we gamble at each step, is the total stack we have at step , and we are interested in when there will be a return or \text{equalization}, i.e., what is the probability that .
Example. [Random walks on the real line]
We shall first consider the simplest non-trivial case of a random walk in , namely the case where the common distribution function of the random variables is given by
We note that in this situation, all paths of length have the same probability, namely .
We can plot the random walk in the plane, where the horizontal axis represents time and the vertical axis represents the value of . Given a sequence of partial sums, we first plot the points , and then for each , we connect and with a straight line segment. The **length** of a path is just the difference in the time values of the beginning and ending points on the path. This can be visualized in Figure .
Fig.1: A random walk .
Returns and First Returns
This subsection studies the random walk in . The p.m.f. of is given by .
We say that an equalization has occurred, or there is a return to the origin at time , if . Note that the returns are possible only after even number of steps. To calculate
the probability of an equalization at time , we need only count the number of paths of length which begin and end at the origin. The number of such paths is clearly
Since each path has probability , we have the following theorem.
Theorem.
The probability of a return to the origin at time is given by
The probability of a return to the origin at an odd time is 0.
A random walk is said to have a first return to the origin at time if , and for all . We define to be the probability of this event. (We also define .) One can think
of the expression as the number of paths of length between the points and that do not touch the horizontal axis except at the endpoints. We have the following theorem.
Theorem.
For , the probabilities and are related by the equation
where and as convention.
There are paths of length which have endpoints and . The collection of such paths can be partitioned into sets, depending upon the time of the first return to the origin. A path in this collection which has a first return to the origin at time consists of an initial segment from to , in which no interior points are on the horizontal axis, and a terminal segment from
to , with no further restrictions on this segment. Thus, the number of paths in the collection which have a first return to the origin at time is given by
If we sum over , we obtain the equation
Dividing both sides of this equation by completes the proof.
◼
Using the generating functions, we have following theorem.
Theorem.
For , the probability of a first return to the origin at time
is given by
We begin by defining the generating functions
and
Note that two sequences converge when . The preceding theorem says that
(The presence of the on the right-hand side is due to the fact that is defined to be , but the last theorem only holds for .) We can solve the above equation for , obtaining
Now, if we can find a closed-form expression for the function , we will also have a closed-form expression for . From the first theorem in this notes, we have
From series expansion, we find that
Therefore, we have
Although it is possible to compute the value of f2m using the Binomial Theorem, it is easier to note that , so that the coefficients can be found by integrating the series for . We obtain, for ,
since
This completes the proof of the theorem.
◼
Probability of Eventual Return
In the symmetric random walk process in , we first examine this question in the case that , and then we consider the general case.
We should be careful when dealing with eventual return since the sample space seems to be the set of all walks of infinite length, and this set is non-denumerable. To avoid difficulties, we will define to be the probability that a first return has occurred no later than time . Thus,
concerns the sample space of all walks of length , which is a finite set. Then it is reasonable to define
This limit clearly exists and is at most one, since the sequence is an increasing sequence, and all of its terms are at most 1. In terms of the probabilities, we see that
Thus
In previous proof we introduced the the generating function
which is convergent for . In fact, it also converges for . Hence we see that
Thus, with probability one, the particle returns to the origin in random walk.
Now we consider the eventual return in . We define to be the probability that the first return to the origin in occurs at time . The quantity is defined in a similar manner. From the preceding theorem, we have
We continue to generalize previous work by defining
Then we see that
These functions will always converge in the interval . In fact, since
for all , the series for converges at as well, and is left continuous at , i.e.,
Thus, we have
so to determine , , it suffices to determine
We let denote this limit. Then we have
In random walk, we have
Using Stirling’s Formula
we have
From this it follows easily that
diverges, so , i.e., in , the probability of an eventual return is 1.
In , we have
Note that converges, so is strictly less than one. This means that in , the probability of an eventual return to the origin is strictly less than one.
Gambler’s Ruin
In this section, we remove the assumption that the random walk is symmetric. Instead, we assume that and are non-negative real numbers with , and that the common distribution function of the jumps of the random walk is
We formulate the problem as follows.
A gambler starts with a “stake” of size . She plays until her capital reaches the value or the value . In the language of Markov chains, these two values correspond to absorbing states. We are interested in studying the probability of occurrence of each of these two outcomes.
We begin by defining to be the probability that the gambler’s stake reaches 0, i.e., she is ruined, before it reaches , given that the initial stake is . We note that and . The fundamental relationship among the ’s is the following:
where . This holds because if her stake equals , and she plays one game, then her stake becomes with probability and with probability . Since , we can rewrite the equation as
From this equation, it is easy to see that
We now use telescoping sums to obtain an equation in which the only unknown is .
Then we can solve for , :