Skip to main content

Random Walks

  • Chapter
  • First Online:
Understanding Markov Chains

Part of the book series: Springer Undergraduate Mathematics Series ((SUMS))

  • 6584 Accesses

Abstract

In this chapter we consider our second important example of discrete-time stochastic process, which is a random walk allowed to evolve over the set \({\mathbb Z}\) of signed integers without any boundary restriction. Of particular importance are the probabilities of return to a given state in finite time, as well as the corresponding mean return time.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
CHF34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
CHF 24.95
Price includes VAT (Switzerland)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
CHF 38.00
Price excludes VAT (Switzerland)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
CHF 47.50
Price excludes VAT (Switzerland)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Note that \(S_{2n}\) is always an even number after we start from \(S_0=0\).

  2. 2.

    Recall that the notation “\(\inf \)” stands for “infimum”, meaning the smallest \(n\ge 0\) such that \(S_n=0\), with \(T_0^r = \infty \) if no such \(n\ge 0\) exists.

  3. 3.

    We used the formula \((1+x)^\alpha = \displaystyle \sum _{k=0}^\infty \frac{x^k}{k!} \alpha ( \alpha - 1 ) \times \cdots \times ( \alpha - (k-1)) \), cf. Relation (A.8).

  4. 4.

    “Any good theorem should have several proofs, the more the better. For two reasons: usually, different proofs have different strengths and weaknesses, and they generalise in different directions - they are not just repetitions of each other. Some of them are good for this application, some are good for that application. They all shed light on the area. If you cannot look at a problem from different directions, it is probably not very interesting; the more perspectives, the better !” - Sir Michael Atiyah.

  5. 5.

    Note that the summation \(\sum _{k=1}^\infty = \sum _{1\le k < \infty }\) actually excludes the value \(k=\infty \).

  6. 6.

    We use the convention \(\infty \times 0 = 0\).

  7. 7.

    Independent and identically distributed.

  8. 8.

    The meaning of \(f(n)\simeq _{_{n \rightarrow \infty }} g(n)\) is \(\lim _{n\rightarrow \infty } f(n)/g(n) = 1\), provided that \(g(n) \not = 0\), \(n\ge 1\).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nicolas Privault .

Exercises

Exercises

Exercise 3.1

We consider the simple random walk \((S_n)_{n\in \mathbb N}\) of Sect. 3.1 with independent increments and started at \(S_0=0\), in which the probability of advance is p and the probability of retreat is \(1-p\).

  1. (a)

    Enumerate all possible sample paths that conduct to \(S_4 = 2\) starting from \(S_0=0\).

  2. (b)

    Show that

    $$ {\mathbb {P}}( S_4 = 2 \mid S_0 = 0 ) = \left( {\begin{array}{c}4\\ 3\end{array}}\right) p^3 (1-p) = \left( {\begin{array}{c}4\\ 1\end{array}}\right) p^3 (1-p) . $$
  3. (c)

    Show that we have

    $$\begin{aligned}&{ {\mathbb {P}}( S_n = k \mid S_0 = 0 ) }\nonumber \\&= \left\{ \begin{array}{ll} \displaystyle \left( {\begin{array}{c}n\\ (n+k)/2\end{array}}\right) p^{(n+k)/2} (1-p)^{(n-k)/2}, &{} n+k \ \text{ even } \text{ and } |k|\le n,\\ {\mathbb {P}}( S_n = k \mid S_0 = 0 ) = 0, &{} n+k \ \text{ odd } \text{ or } |k| > n. \end{array} \right. \end{aligned}$$
    (3.4.24)
  4. (d)

    Show, by a direct argument using a “last step” analysis at time \(n+1\) on random walks, that \(p_{n, k}: = {\mathbb {P}}( S_n = k \mid S_0 = 0 )\) satisfies the difference equation

    $$\begin{aligned} p_{n+1,k} = p p_{n, k-1} + q p_{n, k+1}, \end{aligned}$$
    (3.4.25)

    under the boundary conditions \(p_{0,0}=1\) and \(p_{0,k}=0\), \(k\not =0\).

  5. (e)

    Confirm that \(p_{n, k} = {\mathbb {P}}( S_n = k \mid S_0 = 0 )\) given by (3.4.24) satisfies the equation (3.4.25) and its boundary conditions.

Exercise 3.2

Consider a random walk \((S_n)_{n\in \mathbb N}\) on \({\mathbb Z}\) with independent increments and probabilities p, resp. \(q=1-p\) of moving up by one step, resp. down by one step. Let

$$ T_0 = \inf \{ n \ge 0 \ : \ S_n = 0 \} $$

denote the hitting time of state .

  1. (a)

    Explain why for any \(k\ge 1\) we have

    $$ {\mathrm{I}}\!{\mathrm{E}}[ T_0 \mid S_0 = k ] = k {\mathrm{I}}\!{\mathrm{E}}[ T_0 \mid S_0 = 1 ], $$

    and compute \({\mathrm{I}}\!{\mathrm{E}}[ T_0 \mid S_0 = 1 ]\) using first step analysis when \(q>p\). What can we conclude when \(p\ge q\)?

  2. (b)

    Explain why, by the Markov property, we have

    $$\begin{aligned} {\mathbb {P}}( T_0< \infty \mid S_0 = k ) = \left( {\mathbb {P}}( T_0 < \infty \mid S_0 = 1 ) \right) ^k, \qquad k\ge 1. \end{aligned}$$
  3. (c)

    Using first step analysis for random walks, show that \(\alpha : = {\mathbb {P}}( T_0 < \infty \mid S_0 = 1 )\) satisfies the quadratic equation

    $$ p \alpha ^2 - \alpha + q = p ( \alpha - q/p) ( \alpha - 1) = 0, $$

    and give the values of \({\mathbb {P}}( T_0 < \infty \mid S_0 = 1 )\) and \({\mathbb {P}}( T_0 = \infty \mid S_0 = 1 )\) in the cases \(p < q\) and \(p \ge q\) respectively.

Exercise 3.3

Consider the random walk

$$ S_n : = X_1+\cdots + X_n, \qquad n \ge 1, $$

with \(S_0=0\), where \((X_k)_{k\ge 1}\) is a sequence of Bernoulli random variables with

$$ {\mathbb {P}}( X_k = 1 ) = p\in (0,1), \qquad {\mathbb {P}}( X_k = -1 ) = q \in (0,1), $$

and \(p+q=1\). Recall that the probability generating function (PGF)

$$\begin{aligned} G_{T_0^r} (s) = \sum _{k=0}^\infty s^k {\mathbb {P}}( T_0^r = k ), \qquad s\in [-1,1], \end{aligned}$$
(3.4.26)

of the first return time \(T_0^r : = \inf \{ S_n = 0 \ : \ n \ge 1\}\) to state is given by

$$\begin{aligned} G_{T_0^r} (s) = 1 - \sqrt{1-4pqs^2}, \qquad s\in [-1,1]. \end{aligned}$$
(3.4.27)
  1. (a)

    Compute \({\mathbb {P}}(T_0^r = 0)\) and \({\mathbb {P}}( T_0^r < \infty )\) from \(G_{T_0^r}\).

  2. (b)

    By differentiation of (3.4.26) and (3.4.27), compute \({\mathbb {P}}(T_0^r = 1)\), \({\mathbb {P}}(T_0^r = 2)\), \({\mathbb {P}}(T_0^r = 3)\) and \({\mathbb {P}}(T_0^r = 4)\) using the PGF \(G_{T_0^r}\).

  3. (c)

    Compute \({\mathrm{I}}\!{\mathrm{E}}[ T_0^r \mid T_0^r < \infty ]\) using the PGF \(G_{T_0^r}\).

Exercise 3.4

Consider a simple random walk \((X_n)_{n\ge 0}\) on \({\mathbb Z}\) with respective probabilities p and q of increment and decrement. Let

$$ T_0 : = \inf \{ n \ge 0 \ : \ X_n = 0 \} $$

denote the first hitting time of state , and consider the probability generating function

$$ G_i(s) : = {\mathrm{I}}\!{\mathrm{E}}\big [ s^{T_0} \mid X_0 = i \big ] , \qquad - 1< s < 1, \quad i \in {\mathbb Z}. $$
  1. (a)

    By a first step analysis argument, find the finite difference equation satisfied by \(G_i(s)\), and its boundary condition(s) at \(i=0\) and \(i=\pm \infty \).

  2. (b)

    Find the value of \(G_i(s)\) for all \(i\in {\mathbb Z}\) and \(s\in (0,1)\), and recover the result of (3.4.23) on the probability generating function of the hitting time \(T_0\) of starting from state .

  3. (c)

    Recover Relation (2.2.13)–(3.4.16) using \(G_i (s)\).

  4. (d)

    Recover Relation (2.3.12)–(3.4.18) by differentiation of \(s \mapsto G_i (s)\).

  5. (e)

    Recover the result of (3.4.9) on the probability generating function of the return time \(T_0^r\) to .

Exercise 3.5

Using the probability distribution (3.4.21) of \(T_0^r\), recover the fact that \({\mathrm{I}}\!{\mathrm{E}}[ T_0^r \mathbbm {1}_{\{ T_0^r < \infty \} } \mid S_0 = 0 ] =\infty \), when \(p=q=1/2\).

Exercise 3.6

Consider a sequence \((X_k)_{k \ge 1}\) of independent Bernoulli random variables with

$$ {\mathbb {P}}( X_k = 1 ) = p, \quad \text{ and } \quad {\mathbb {P}}( X_k = -1 ) = q, \qquad k\ge 1, $$

where \(p+q=1\), and let the process \((M_n)_{n\in \mathbb N}\) be defined by \(M_0 := 0\) and

$$ M_n := \sum _{k=1}^n 2^{k-1} X_k, \qquad n \ge 1. $$
  1. (a)

    Compute \({\mathrm{I}}\!{\mathrm{E}}[M_n]\) for all \(n\ge 0\).

  2. (b)

    Consider the hitting time \( \tau : = \inf \{ n \ge 1 \ : \ M_n = 1 \} \) and the stopped process

    $$ M_{\min ( n , \tau ) } = M_n \mathbbm {1}_{ \{ n < \tau \} } + \mathbbm {1}_{ \{ \tau \le n\} }, \qquad n \in \mathbb N. $$

    Determine the possible values of \(M_{\min ( n , \tau )}\), and the probability distribution of \(M_{\min ( n , \tau )}\) at any time \(n\ge 1\).

  3. (c)

    Give an interpretation of the stopped process \((M_{\min ( n , \tau ) })_{n\in \mathbb N}\) in terms of strategy in a game started at \(M_0=0\).

  4. (d)

    Based on the result of part (b), compute \({\mathrm{I}}\!{\mathrm{E}}[ M_{\min ( n , \tau ) } ]\) for all \(n \ge 1\).

Exercise 3.7

Winning streaks. Consider a sequence \((X_n)_{n\ge 1}\) of independent Bernoulli random variables with the distribution

$$ {\mathbb {P}}( X_n = 1 ) = p, \qquad {\mathbb {P}}( X_n = 0 ) = q, \qquad n \ge 1, $$

with \(q:=1-p\). For some \(m\ge 1\), let \(T^{(m)}\) denote the time of the first appearance of m consecutive “1” in the sequence \((X_n)_{n\ge 1}\). For example, for \(m=4\) the following sequence

$$ ( \overset{1 \atop \downarrow }{0},\overset{2 \atop \downarrow }{1},\overset{3 \atop \downarrow }{1},\overset{4 \atop \downarrow }{0}, \underset{ 4 \text{ times } }{\underbrace{ \overset{5 \atop \downarrow }{1} , \overset{6 \atop \downarrow }{1}, \overset{7 \atop \downarrow }{1}, \overset{8 \atop \downarrow }{1} }} , 0 , 1, 1, 0, \ldots ) $$

yields \(T^{(4)} = 8\).

  1. (a)

    Compute \({\mathbb {P}}(T^{(m)} < m)\), \({\mathbb {P}}(T^{(m)} = m)\), \({\mathbb {P}}(T^{(m)} = m+1)\), and \({\mathbb {P}}(T^{(m)} = m+2)\).

  2. (b)

    Show that the probability generating function

    $$ G_{T^{(m)}}(s) : = {\mathrm{I}}\!{\mathrm{E}}\left[ s^{T^{(m)}} \mathbbm {1}_{\{ T^{(m)} < \infty \} } \right] , \qquad s\in (-1,1), $$

    satisfies

    $$\begin{aligned} G_{T^{(m)}} (s) = p^m s^m + \sum _{k=0}^{m-1} p^k q s^{k+1} G_{T^{(m)}} (s) , \qquad s\in (-1,1). \end{aligned}$$
    (3.4.28)

    Hint: Look successively at all possible starting patterns of the form

    $$ (1, \ldots , 1, \overset{ k \atop \downarrow }{1}, 0 , \ldots ), $$

    where \(k=0,1,\ldots , m\), compute their respective probabilities, and apply a “k-step analysis” argument.

  3. (c)

    From (3.4.28), compute the probability generating function \(G_{T^{(m)}}\) of \(T^{(m)}\) for all \(s\in (-1,1)\).

    Hint: We have

    $$ \sum _{k=0}^{m-1} x^k = \frac{1-x^m}{1-x}, \qquad x\in (-1,1). $$
  4. (d)

    From the probability generating function \(G_{T^{(m)}}(s)\), compute \({\mathrm{I}}\!{\mathrm{E}}[ T^{(m)}]\) for all \(m\ge 1\).

    Hint: It can be simpler to differentiate inside (3.4.28) and to use the relation

    $$ (1-x) \sum _{k=0}^{m-1} (k+1)x^k + m x^m = \frac{1-x^m}{1-x}, \qquad x\in (-1,1). $$

Exercise 3.8

Consider a random walk \((S_n)_{n\in \mathbb N}\) on \({\mathbb Z}\) with increments \(\pm 1\), started at \(S_0 = 0\). Recall that the number of paths joining states and over 2m time steps is

$$\begin{aligned} \left( {\begin{array}{c}2m\\ m+k\end{array}}\right) . \end{aligned}$$
(3.4.29)
  1. (a)

    Compute the total number of paths joining \(S_1=1\) to \(S_{2n-1}=1\).

    Hint: Apply the formula (3.4.29).

  2. (b)

    Compute the total number of paths joining \(S_1=1\) to \(S_{2n-1}=-1\).

    Hint: Apply the formula (3.4.29).

  3. (c)

    Show that to every one path joining \(S_1=1\) to \(S_{2n-1}=1\) by crossing or hitting we can associate one path joining \(S_1=1\) to \(S_{2n-1}=-1\), in a one-to-one correspondence. Hint: Draw a sample path joining \(S_1=1\) to \(S_{2n-1}=1\), and reflect it in such a way that the reflected path then joins \(S_1=1\) to \(S_{2n-1}=-1\).

  4. (d)

    Compute the total number of paths joining \(S_1=1\) to \(S_{2n-1}=1\) by crossing or hitting . Hint: Combine the answers to part (b) and part (c).

  5. (e)

    Compute the total number of paths joining \(S_1=1\) to \(S_{2n-1}=1\) without crossing or hitting . Hint: Combine the answers to part (a) and (d).

  6. (f)

    Give the total number of paths joining \(S_0=0\) to \(S_{2n}=0\) without crossing or hitting between time 1 and time \(2n-1\). Hint: Apply two times the answer to part (e). A drawing is recommended.

Exercise 3.9

Range process. Consider the random walk \((S_n)_{n\ge 0}\) defined by \(S_0=0\) and

$$ S_n: = X_1 + \cdots + X_n, \qquad n \ge 1, $$

where \((X_k)_{k\ge 1}\) is an i.i.d.Footnote 7 family of \(\{ - 1 , + 1 \}\)-valued random variables with distribution

$$ \left\{ \begin{array}{l} {\mathbb {P}}( X_k = +1 ) = p,\\ {\mathbb {P}}( X_k = -1 ) = q, \end{array} \right. $$

\(k \ge 1\), where \(p+q=1\). We let \(R_n\) denote the range of \((S_0,S_1,\ldots , S_n)\), i.e. the (random) number of distinct values appearing in the sequence \((S_0,S_1,\ldots , S_n)\).

  1. (a)

    Explain why

    $$ R_n = 1 + \left( \sup _{k=0,1,\ldots , n} S_k \right) - \left( \inf _{k=0,1,\ldots , n} S_k \right) , $$

    and give the value of \(R_0\) and \(R_1\).

  2. (b)

    Show that for all \(k\ge 1\), \(R_k-R_{k-1}\) is a Bernoulli random variable, and that

    $$ {\mathbb {P}}( R_k - R_{k-1} = 1 ) = {\mathbb {P}}( S_k - S_0 \not = 0, S_k - S_1 \not = 0, \ldots , S_k - S_{k-1} \not = 0 ) . $$
  3. (c)

    Show that for all \(k\ge 1\) we have

    $$ {\mathbb {P}}( R_k - R_{k-1} = 1 ) = {\mathbb {P}}( X_1 \not = 0, X_1 + X_2 \not = 0, \ldots , X_1+\cdots +X_k \not = 0 ) . $$
  4. (d)

    Show why the telescoping identity \(\displaystyle R_n = R_0 + \sum _{k=1}^n ( R_k - R_{k-1} )\) holds for all \(n \in \mathbb N\).

  5. (e)

    Show that \( {\mathbb {P}}( T_0^r = \infty ) = \lim _{k \rightarrow \infty } {\mathbb {P}}( T_0^r > k ) \).

  6. (f)

    From the results of Questions (c) and (d), show that

    $$ {\mathrm{I}}\!{\mathrm{E}}[ R_n ] = \sum _{k=0}^n {\mathbb {P}}( T_0^r > k ) , \qquad n \in \mathbb N, $$

    where \( T_0^r = \inf \{ n \ge 1 \ : \ S_n = 0 \} \) is the time of first return to of the random walk.

  7. (g)

    From the results of Questions (e) and (f), show that

    $$ {\mathbb {P}}( T_0^r = \infty ) = \lim _{n\rightarrow \infty } \frac{1}{n} {\mathrm{I}}\!{\mathrm{E}}[ R_n ] . $$
  8. (h)

    Show that

    $$ \lim _{n\rightarrow \infty } \frac{1}{n} {\mathrm{I}}\!{\mathrm{E}}[ R_n ] = 0. $$

    when \(p=1/2\), and that \({\mathrm{I}}\!{\mathrm{E}}[ R_n ] \simeq _{_{n\rightarrow \infty }} n | p - q |\), when \(p \not = 1/2\).Footnote 8

Fig. 3.5
figure 5

Illustration of the range process

In Fig. 3.5 the height at time n of the colored area coincides with \(R_n-1\).

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Singapore Pte Ltd.

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Privault, N. (2018). Random Walks. In: Understanding Markov Chains. Springer Undergraduate Mathematics Series. Springer, Singapore. https://doi.org/10.1007/978-981-13-0659-4_3

Download citation

Publish with us

Policies and ethics

  NODES
Note 9