Problems

Filters
Clear Filters

71 problems found

2014 Paper 2 Q13
D: 1600.0 B: 1469.5

A random number generator prints out a sequence of integers \(I_1, I_2, I_3, \dots\). Each integer is independently equally likely to be any one of \(1, 2, \dots, n\), where \(n\) is fixed. The random variable \(X\) takes the value \(r\), where \(I_r\) is the first integer which is a repeat of some earlier integer. Write down an expression for \(\mathbb{P}(X=4)\).

  1. Find an expression for \(\mathbb{P}(X=r)\), where \(2\le r\le n+1\). Hence show that, for any positive integer \(n\), \[ \frac 1n + \left(1-\frac1n\right) \frac 2 n + \left(1-\frac1n\right)\left(1-\frac2n\right) \frac3 n + \cdots \ = \ 1 \,. \]
  2. Write down an expression for \(\mathbb{E}(X)\). (You do not need to simplify it.)
  3. Write down an expression for \(\mathbb{P}(X\ge k)\).
  4. Show that, for any discrete random variable \(Y\) taking the values \(1, 2, \dots, N\), \[ \mathbb{E}(Y) = \sum_{k=1}^N \mathbb{P}(Y\ge k)\,. \] Hence show that, for any positive integer \(n\), \[ \left(1-\frac{1^2}n\right) + \left(1-\frac1n\right)\left(1-\frac{2^2}n\right) + \left(1-\frac1n\right)\left(1-\frac{2}n\right)\left(1-\frac{3^2}n\right) + \cdots \ = \ 0. \]


Solution: \begin{align*} && \mathbb{P}(X > 4) &= 1 \cdot \frac{n-1}{n} \cdot \frac{n-2}{n} \cdot \frac{n-3}{n} \\ && \mathbb{P}(X > 3) &= 1 \cdot \frac{n-1}{n} \cdot \frac{n-2}{n} \\ \Rightarrow && \mathbb{P}(X =4) &= \mathbb{P}(X > 3) - \mathbb{P}(X > 4) \\ &&&= \frac{(n-1)(n-2)}{n^2} \left (1 - \frac{n-3}{n} \right) \\ &&&= \frac{3(n-1)(n-2)}{n^3} \end{align*}

  1. Notice that \begin{align*} && \mathbb{P}(X > r) &= \frac{n-1}{n} \cdots \frac{n-r+1}{n} \\ \Rightarrow && \mathbb{P}(X = r) &= \frac{n-1}{n} \cdots \frac{n-r+2}{n} \left (1 - \frac{n-r+1}{n} \right) \\ &&&= \frac{(n-1)\cdots(n-r+2)(r-1)}{n^{r-1}} \\ &&&= \left (1 - \frac{1}n \right)\left (1 - \frac{2}{n} \right) \cdots \left (1 - \frac{r-2}{n} \right) \frac{r-1}{n} \\ \Rightarrow && 1 &= \sum \mathbb{P}(X = r) \\ &&&= \sum_{r=2}^{n+1} \mathbb{P}(X = r) \\ &&&= \frac 1n + \left(1-\frac1n\right) \frac 2 n + \left(1-\frac1n\right)\left(1-\frac2n\right) \frac3 n + \cdots \end{align*}
  2. \(\,\) \begin{align*} \mathbb{E}(X) &= \sum_{r=2}^{n+1} r\cdot\mathbb{P}(X = r) \\ &= \frac 2n + \left(1-\frac1n\right) \frac {2\cdot3} n + \left(1-\frac1n\right)\left(1-\frac2n\right) \frac{3\cdot4} n + \cdots \end{align*}
  3. \(\displaystyle \mathbb{P}(X \geq k) = \frac{n-1}{n} \cdots \frac{n-r+2}{n}\)
  4. \(\,\) \begin{align*} && \mathbb{E}(Y) &= \sum_{r=1}^N r \cdot \mathbb{P}(Y = r) \\ &&&= \sum_{r=1}^N \sum_{j=1}^r \mathbb{P}(Y = r) \\ &&&= \sum_{j=1}^N \sum_{r=j}^N \mathbb{P}(Y=r) \\ &&&= \sum_{j=1}^N \mathbb{P}(Y \geq j) \end{align*} Let \(P_k = \left(1-\frac1n\right)\left(1-\frac2n\right) \cdots \left(1-\frac1n\right)\left(1-\frac{k}n\right) \) \begin{align*} && \mathbb{E}(X) &= P_1 \frac{1 \cdot 2 }{n} + P_2 \cdot \frac{2 \cdot 3}{n} + \cdots + P_k \cdot \frac{k(k+1)}{n} + \cdots \\ && &= \sum_{k=1}^{n} \frac{k^2}{n}P_k + \sum_{k=1}^{n} \frac{k}{n}P_k \\ && \text{Using the identity } & \frac{k}{n}P_k = \frac{k}{n} \prod_{i=1}^{k-1} \left(1-\frac{i}{n}\right) = P_k - P_{k+1}: \\ && \sum_{k=1}^{n} \frac{k}{n}P_k &= (P_1 - P_2) + (P_2 - P_3) + \cdots + (P_n - P_{n+1}) \\ && &= P_1 - P_{n+1} = 1 - 0 = 1 \\ \\ \Rightarrow && \mathbb{E}(X) &= \sum_{k=1}^{n} \frac{k^2}{n}P_k + 1 \\ && &= \mathbb{P}(X \geq 1) + \mathbb{P}(X \geq 2) + \mathbb{P}(X \geq 3) + \cdots \\ && &= 1 + P_1 + P_2 + P_3 + \cdots \\ && &= 1 + \sum_{k=1}^{n} P_k \\ \\ \Rightarrow && 1 + \sum_{k=1}^{n} P_k &= \sum_{k=1}^{n} \frac{k^2}{n}P_k + 1 \\ \Rightarrow && \sum_{k=1}^{n} P_k &= \sum_{k=1}^{n} \frac{k^2}{n}P_k \\ \Rightarrow && 0 &= \sum_{k=1}^{n} P_k \left( 1 - \frac{k^2}{n} \right) \end{align*}

2014 Paper 3 Q12
D: 1700.0 B: 1500.0

The random variable \(X\) has probability density function \(f(x)\) (which you may assume is differentiable) and cumulative distribution function \(F(x)\) where \(-\infty < x < \infty \). The random variable \(Y\) is defined by \(Y= \e^X\). You may assume throughout this question that \(X\) and \(Y\) have unique modes.

  1. Find the median value \(y_m\) of \(Y\) in terms of the median value \(x_m\) of \(X\).
  2. Show that the probability density function of \(Y\) is \(f(\ln y)/y\), and deduce that the mode \(\lambda\) of \(Y\) satisfies \(\f'(\ln \lambda) = \f(\ln \lambda)\).
  3. Suppose now that \(X \sim {\rm N} (\mu,\sigma^2)\), so that \[ f(x) = \frac{1}{\sigma \sqrt{2\pi}\,} \e^{-(x-\mu)^2/(2\sigma^2)} \,. \] Explain why \[\frac{1}{\sigma \sqrt{2\pi}\,} \int_{-\infty}^{\infty}\e^{-(x-\mu-\sigma^2)^2/(2\sigma^2)} \d x = 1 \] and hence show that \( \E(Y) = \e ^{\mu+\frac12\sigma^2}\).
  4. Show that, when \(X \sim {\rm N} (\mu,\sigma^2)\), \[ \lambda < y_m < \E(Y)\,. \]


Solution:

  1. \begin{align*} && \frac12 &= \mathbb{P}(X \leq x_m) \\ \Leftrightarrow && \frac12 &= \mathbb{P}(e^X \leq e^{x_m} = y_m) \end{align*} Therefore the median is \(y_m = e^{x_m}\)
  2. \begin{align*} && \mathbb{P}(Y \leq y) &= \mathbb{P}(e^X \leq y) \\ &&&= \mathbb{P}(X \leq \ln y) \\ &&&= F(\ln y) \\ \Rightarrow && f_Y(y) &= f(\ln y)/y \\ \\ && f'_Y(y) &= \frac{f'(\ln y) - f(\ln y)}{y^2} \end{align*} Therefore since the mode satisfies \(f'_Y = 0\) we must have \(f'(\ln \lambda ) = f(\ln \lambda)\)
  3. This is the integral of the pdf of \(N(\mu + \sigma^2, \sigma^2)\) and therefore is clearly \(1\). \begin{align*} && \E[Y] &= \int_{-\infty}^{\infty} e^x \cdot \frac{1}{\sqrt{2\pi \sigma^2}} e^{-(x-\mu)^2/(2\sigma^2)} \d x \\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (x - (x-\mu)^2/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp ((2x \sigma^2- (x-\mu)^2)/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu-\sigma^2)^2+2\mu \sigma^2-\sigma^4)/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu+\sigma^2)^2)/(2\sigma^2)+\mu +\frac12\sigma^2) \d x\\ &&&= \e^{\mu +\frac12\sigma^2}\frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu-\sigma^2)^2)/(2\sigma^2)) \d x\\ &&&= \e^{\mu +\frac12\sigma^2} \end{align*}
  4. Notice that \(y_m = e^\mu < e^{\mu + \tfrac12 \sigma^2} = \E[Y]\), so it suffices to prove that \(\lambda < e^{\mu}\) Notice that \(f'(x) - f(x) = f(x)[-(x-\mu)/\sigma^2 - 1]\) and therefore \(\ln y - \mu = -\sigma^2\) so \(\lambda = e^{\mu - \sigma^2}\) which is clearly less than \(e^{\mu}\) as required.

2014 Paper 3 Q13
D: 1700.0 B: 1500.0

I play a game which has repeated rounds. Before the first round, my score is \(0\). Each round can have three outcomes:

  1. my score is unchanged and the game ends;
  2. my score is unchanged and I continue to the next round;
  3. my score is increased by one and I continue to the next round.
The probabilities of these outcomes are \(a\), \(b\) and \(c\), respectively (the same in each round), where \(a+b+c=1\) and \(abc\ne0\). The random variable \(N\) represents my score at the end of a randomly chosen game. Let \(G(t)\) be the probability generating function of \(N\).
  1. Suppose in the first round, the game ends. Show that the probability generating function conditional on this happening is 1.
  2. Suppose in the first round, the game continues to the next round with no change in score. Show that the probability generating function conditional on this happening is \(G(t)\).
  3. By comparing the coefficients of \(t^n\), show that $ G(t) = a + bG(t) + ctG(t)\,. $ Deduce that, for \(n\ge0\), \[ P(N=n) = \frac{ac^n}{(1-b)^{n+1}}\,. \]
  4. Show further that, for \(n\ge0\), \[ P(N=n) = \frac{\mu^n}{(1+\mu)^{n+1}}\,, \] where \(\mu=\E(N)\).


Solution:

  1. If the game ends in the first round then the score is exactly \(0\) and the pgf is \(1\cdot x^0 = 1\)
  2. If the game moves onto the next round with no change in the first round then it's as if nothing happened, therefore the pgf is the original pgf \(G(t)\)
  3. If the game moves into the next round with the score increased by one, then the pgf is \(tG(t)\) since all the scores are increased by \(1\). Therefore \begin{align*} && G(t) &= \E[t^N] \\ &&&= \E[\E[t^N | \text{first round}]] \\ &&&= a + bG(t) + ctG(t) \\ \Rightarrow && G(t)(1-b-ct) = a \\ \Rightarrow && G(t) &= \frac{a}{(1-b)-ct} \\ &&&= \frac{a}{(1-b)} \frac{1}{1- \left(\frac{c}{1-b}\right)t} \\ &&&= \sum_{n=0}^\infty \frac{a}{1-b} \frac{c^n}{(1-b)^n} t^n\\ &&&= \sum_{n=0}^{\infty} \frac{ac^n}{(1-b)^{n+1}}t^n \end{align*} Therefore by comparing coefficients, \(\mathbb{P}(N=n) = \frac{ac^n}{(1-b)^{n+1}}\)
  4. \(\,\) \begin{align*} && \E[N] &= G'(1) \\ &&&= \frac{ac}{((1-b)-c)^2} \\ &&&= \frac{ac}{a^2} = \frac{c}{a} \\ \\ && \frac{\mu^n}{(1+\mu)^{n+1}} &= \frac{c^na^{-n}}{(a+c)^{n+1}a^{-n-1}} \\ &&&= \frac{ac^n}{(a+c)^{n+1}}\\ &&&= \frac{ac^n}{(1-b)^{n+1}}\\ &&&= \mathbb{P}(N=n) \end{align*} as required

2013 Paper 1 Q13
D: 1516.0 B: 1532.0

From the integers \(1, 2, \ldots , 52\), I choose seven (distinct) integers at random, all choices being equally likely. From these seven, I discard any pair that sum to 53. Let \(X\) be the random variable the value of which is the number of discarded pairs. Find the probability distribution of \(X\) and show that \(\E (X) = \frac 7 {17}\). Note: \(7\times 17 \times 47 =5593\).


Solution: There are \(\binom{26}3\binom{23}{1}2\) ways to obtain \(3\) pairs There are \(\binom{26}2 \binom{24}3 \cdot 2^3\) ways to obtain \(2\) pairs There are \(\binom{26}1 \binom{25}5 \cdot 2^5\) ways to obtain \(1\) pairs There are \(\binom{26}7 \cdot 2^7\) ways to obtain \(0\) pairs There are \(\binom{52}{7}\) ways to choose our integers, so \begin{align*} && \mathbb{P}(X = 3) &= \frac{\binom{26}{3} \cdot \binom{23}{1} \cdot 2}{\binom{52}{7}} \\ &&&= \frac{7! \cdot 26 \cdot 25 \cdot 24 \cdot 23 \cdot 2}{3! \cdot 52 \cdot 51 \cdot 50 \cdot 49 \cdot 48 \cdot 47 \cdot 46} \\ &&&= \frac{7 \cdot 6 \cdot 5 \cdot 4 }{51 \cdot 2\cdot 49 \cdot 2\cdot 47 \cdot 2} \\ &&&= \frac{ 5 }{17\cdot 7\cdot 47} = \frac{5}{5593} \\ \\ && \mathbb{P}(X = 2) &= \frac{\binom{26}2 \binom{24}3 \cdot 2^3}{\binom{52}{7}} \\ &&&= \frac{220}{5593} \\ \\ && \mathbb{P}(X = 1) &= \frac{\binom{26}1 \binom{25}5 \cdot 2^5}{\binom{52}{7}} \\ &&&= \frac{1848}{5593} \\ \\ && \mathbb{P}(X = 0) &= \frac{\binom{26}7 \cdot 2^7}{\binom{52}{7}} \\ &&&= \frac{3520}{5593} \\ \\ && \mathbb{E}(X) &= \frac{1848}{5593} + 2 \cdot \frac{220}{5593} + 3 \cdot \frac{5}{5593} \\ &&&= \frac{2303}{5593} = \frac{7}{17} \end{align*} Notice we can find the expected value directly: Let \(X_i\) be the random variable the \(i\)th number is discarded. Notice that \(\mathbb{E}(X) = \mathbb{E}\left (\frac12 \left (X_1 +X_2 +X_3 +X_4 +X_5 +X_6 +X_7\right) \right)\) and also notice that each \(X_i\) has the same distribution (although not independent!). Then \begin{align*} &&\mathbb{E}(X) &= \frac72 \mathbb{E}(X_i) \\ &&&= \frac72 \cdot \left (1 - \frac{50}{51} \cdot \frac{49}{50} \cdots \frac{45}{46} \right) = \frac74 \left ( 1 - \frac{45}{51}\right) \\ &&&= \frac72 \cdot \frac{6}{51} \\ &&&= \frac7{17} \end{align*}

2013 Paper 2 Q12
D: 1600.0 B: 1484.0

The random variable \(U\) has a Poisson distribution with parameter \(\lambda\). The random variables \(X\) and \(Y\) are defined as follows. \begin{align*} X&= \begin{cases} U & \text{ if \(U\) is 1, 3, 5, 7, \(\ldots\,\)} \\ 0 & \text{ otherwise} \end{cases} \\ Y&= \begin{cases} U & \text{ if \(U\) is 2, 4, 6, 8, \(\ldots\,\) } \\ 0 & \text{ otherwise} \end{cases} \end{align*}

  1. Find \(\E(X)\) and \(\E(Y)\) in terms of \(\lambda\), \(\alpha\) and \(\beta\), where \[ \alpha = 1+\frac{\lambda^2}{2!}+\frac{\lambda^4}{4!} +\cdots\, \text{ \ \ and \ \ } \beta = \frac{\lambda}{1!} + \frac{\lambda^3}{3!} + \frac{\lambda^5}{5!} +\cdots\,. \]
  2. Show that \[ \var(X) = \frac{\lambda\alpha+\lambda^2\beta}{\alpha+\beta} - \frac{\lambda^2\alpha^2}{(\alpha+\beta)^2} \] and obtain the corresponding expression for \(\var(Y)\). Are there any non-zero values of \(\lambda\) for which \( \var(X) + \var(Y) = \var(X+Y)\,\)?


Solution:

  1. \begin{align*} \mathbb{E}(X) &= \sum_{r=1}^\infty r \mathbb{P}(X = r) \\ &= \sum_{j=1}^{\infty} (2j-1)\mathbb{P}(U=2j-1) \\ &= \sum_{j=1}^{\infty}(2j-1) \frac{e^{-\lambda} \lambda^{2j-1}}{(2j-1)!} \\ &= \sum_{j=1}^{\infty} e^{-\lambda} \frac{\lambda^{2j-1}}{(2j-2)!} \\ &= \lambda e^{-\lambda} \sum_{j=1}^{\infty} \frac{\lambda^{2j-2}}{(2j-2)!} \\ &= \lambda e^{-\lambda} \alpha \end{align*} Since \(\mathbb{E}(X+Y) = \lambda, \mathbb{E}(Y) = \lambda(1-e^{-\lambda}\alpha) = \lambda(e^{-\lambda}(\alpha+\beta) - e^{-\lambda}\alpha) = \lambda e^{-\lambda} \beta\). Alternatively, as \(\beta + \alpha = e^{\lambda}\), \(\mathbb{E}(X) = \frac{\lambda \alpha}{\alpha+\beta}, \mathbb{E}(Y) = \frac{\lambda \beta}{\alpha+\beta}\)
  2. \begin{align*} \textrm{Var}(X) &= \mathbb{E}(X^2) - [\mathbb{E}(X) ]^2 \\ &= \sum_{odd} r^2 \mathbb{P}(U = r) - \left [ \mathbb{E}(X) \right]^2 \\ &= \sum_{odd} (r(r-1)+r)\frac{e^{-\lambda}\lambda^r}{r!} - \frac{\lambda^2 \alpha^2}{(\alpha+\beta)^2} \\ &= \sum_{odd} \frac{e^{-\lambda}\lambda^r}{(r-2)!}+\sum_{odd} \frac{e^{-\lambda}\lambda^r}{(r-1)!} - \frac{\lambda^2 \alpha^2}{(\alpha+\beta)^2} \\ &= e^{-\lambda}\lambda^2 \beta + e^{-\lambda}\lambda \alpha - \frac{\lambda^2 \alpha^2}{(\alpha+\beta)^2} \\ &= \frac{\lambda \alpha + \lambda^2 \beta}{\alpha+\beta}- \frac{\lambda^2 \alpha^2}{(\alpha+\beta)^2} \end{align*} Similarly, \begin{align*} \textrm{Var}(Y) &= \mathbb{E}(Y^2) - [\mathbb{E}(Y) ]^2 \\ &= \sum_{even} r^2 \mathbb{P}(U = r) - \left [ \mathbb{E}(Y) \right]^2 \\ &= \sum_{even} (r(r-1)+r)\frac{e^{-\lambda}\lambda^r}{r!} - \frac{\lambda^2 \beta^2}{(\alpha+\beta)^2} \\ &= e^{-\lambda}\lambda^2\alpha + e^{-\lambda}\lambda \beta - \frac{\lambda^2 \beta^2}{(\alpha+\beta)^2} \\ &= \frac{\lambda \beta + \lambda^2 \alpha}{\alpha+\beta}- \frac{\lambda^2 \beta^2}{(\alpha+\beta)^2} \end{align*} Since \(\textrm{Var}(X+Y) = \textrm{Var}(U) = \lambda\), we are interested in solving: \begin{align*} \lambda &= \frac{\lambda \alpha + \lambda^2 \beta}{\alpha+\beta}- \frac{\lambda^2 \alpha^2}{(\alpha+\beta)^2} + \frac{\lambda \beta + \lambda^2 \alpha}{\alpha+\beta}- \frac{\lambda^2 \beta^2}{(\alpha+\beta)^2} \\ &= \frac{\lambda(\alpha+\beta) + \lambda^2(\alpha+\beta)}{\alpha+\beta} - \frac{\lambda^2(\alpha^2+\beta^2)}{(\alpha+\beta)^2} \\ &= \lambda + \lambda^2 \frac{(\alpha+\beta)^2 - (\alpha^2+\beta^2)}{(\alpha+\beta)^2} \\ &= \lambda + \lambda^2 \frac{2\alpha\beta}{(\alpha+\beta)^2} \end{align*} which is clearly not possible if \(\lambda \neq 0\)

2013 Paper 3 Q12
D: 1700.0 B: 1500.0

A list consists only of letters \(A\) and \(B\) arranged in a row. In the list, there are \(a\) letter \(A\)s and \(b\) letter \(B\)s, where \(a\ge2\) and \(b\ge2\), and \(a+b=n\). Each possible ordering of the letters is equally probable. The random variable \(X_1\) is defined by \[ X_1 = \begin{cases} 1 & \text{if the first letter in the row is \(A\)};\\ 0 & \text{otherwise.} \end{cases} \] The random variables \(X_k\) (\(2 \le k \le n\)) are defined by \[ X_k = \begin{cases} 1 & \text{if the \((k-1)\)th letter is \(B\) and the \(k\)th is \(A\)};\\ 0 & \text{otherwise.} \end{cases} \] The random variable \(S\) is defined by \(S = \sum\limits_ {i=1}^n X_i\,\).

  1. Find expressions for \(\E(X_i)\), distinguishing between the cases \(i=1\) and \(i\ne1\), and show that \(\E(S)= \dfrac{a(b+1)}n\,\).
  2. Show that:
    1. for \(j\ge3\), \(\E(X_1X_j) = \dfrac{a(a-1)b}{n(n-1)(n-2)}\,\);
    2. \[ \sum\limits_{i=2}^{n-2} \bigg( \sum\limits_{j=i+2}^n \E(X_iX_j)\bigg) = \dfrac{a(a-1)b(b-1)}{2n(n-1)}\,\]
    3. \(\var(S) = \dfrac {a(a-1)b(b+1)}{n^2(n-1)}\,\).


Solution:

  1. Notice that \(\E[X_1] = \frac{a}{n}\) and consider \(\E[X_i]\) with \(i > 1\). the probability that this is \(1\) is \(\frac{b}{n} \cdot \frac{a}{n-1}\). So \begin{align*} && \E[S] &= \E[X_1] + \sum_{i=2}^n \E[X_i] \\ &&&= \frac{a}{n} + (n-1) \frac{ab}{n(n-1)} \\ &&&= \frac{a(b+1)}{n} \end{align*}
    1. The probability \(X_1X_j = 1\) is \(\frac{a}{n} \cdot \frac{b}{n-1} \cdot \frac{a-1}{n-2} = \frac{a(a-1)b}{n(n-1)(n-2)}\) since there is nothing special about the order, and the first is an \(A\) with probability \(\frac{a}{n}\) and given this occurs there are now \(a-1\) \(A\) and \(n-1\) letters left etc... Therefore \(\E[X_1X_j] = \frac{a(a-1)b}{n(n-1)(n-2)}\)
    2. \(\E[X_iX_j]\) when the pairs don't overlap is \(\frac{a}{n} \frac{b}{n-1} \frac{a-1}{n-2} \frac{b-1}{n-3}\), and so \begin{align*} && \sum\limits_{i=2}^{n-2} \bigg( \sum\limits_{j=i+2}^n \E(X_iX_j)\bigg) &= \sum\limits_{i=2}^{n-2} \bigg( \sum\limits_{j=i+2}^n \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)}\bigg) \\ &&&= \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)}\sum\limits_{i=2}^{n-2} \bigg( \sum\limits_{j=i+2}^n 1\bigg) \\ &&&= \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)}\sum\limits_{i=2}^{n-2} (n-(i+1)) \\ &&&= \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)} \left ((n-1)(n-3)-\frac{(n-2)(n-1)}{2}+1 \right) \\ &&&= \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)} \left ( \frac{2n^2-8n-6-n^2+3n-2+2}{2}\right) \\ &&&= \frac{a(a-1)b(b-1)}{n(n-1)(n-2)(n-3)} \left ( \frac{n^2-5n-6}{2}\right) \\ &&&= \frac{a(a-1)b(b-1)}{2n(n-1)} \end{align*}
    3. We also need to consider the other cross terms. \(X_iX_{i+1}=0\). (Since \(X_i = 1\) means the \(i\)th letter is \(A\) and \(X_{i+1} = 1\) means the \(i\)th letter is \(B\)). It's the same story for \(X_1X_2\), and so all the cross terms are accounted for. Therefore \begin{align*} && \E[S^2] &= \E \left [\sum X_i^2 + 2\sum_{i \neq j} X_i X_j \right] \\ &&&= \frac{a(b+1)}{n} +2(n-2)\frac{a(a-1)b}{n(n-1)(n-2)}+ 2 \frac{a(a-1)b(b-1)}{2n(n-1)} \\ &&&= \frac{a(b+1)}{n} +\frac{2a(a-1)b}{n(n-1)} + \frac{a(a-1)b(b-1)}{n(n-1)} \\ &&&= \frac{a(b+1)}{n} +\frac{a(a-1)b(b+1)}{n(n-1)} \\ && \var[S] &= \E[S^2] - \left ( \E[S] \right)^2 \\ &&&= \frac{a(b+1)}{n} + \frac{a(a-1)b(b+1)}{n(n-1)} - \frac{a^2(b+1)^2}{n^2} \\ &&&= \frac{a(b+1) \left (n(n-1) + (a-1)b n -a(b+1)(n-1) \right)}{n^2(n-1)} \\ &&&= \frac{a(b+1) \left ( (n-a)(n-b-1) \right)}{n^2(n-1)} \\ &&&= \frac{a(b+1) \left ( b(a-1) \right)}{n^2(n-1)} \\ \end{align*}

2013 Paper 3 Q13
D: 1700.0 B: 1484.0

  1. The continuous random variable \(X\) satisfies \(0\le X\le 1\), and has probability density function \(\f(x)\) and cumulative distribution function \(\F(x)\). The greatest value of \(\f(x)\) is \(M\), so that \(0\le \f(x) \le M\).
    1. Show that \(0\le \F(x) \le Mx\) for \(0\le x\le1\).
    2. For any function \(\g(x)\), show that \[ \int_0^1 2 \g(x) \F(x) \f(x) \d x = \g(1) - \int_0^1 \g'(x) \big( \F(x)\big)^2 \d x \,. \]
  2. The continuous random variable \(Y\) satisfies \(0\le Y\le 1\), and has probability density function \(k \F(y) \f(y)\), where \(\f\) and \(\F\) are as above.
    1. Determine the value of the constant \(k\).
    2. Show that \[ 1+ \frac{nM}{n+1}\mu_{n+1} - \frac{nM}{n+1} \le \E(Y^n) \le 2M\mu_{n+1}\,, \] where \(\mu_{n+1} = \E(X^{n+1})\) and \(n\ge0\).
    3. Hence show that, for \(n\ge 1\), \[ \mu _n \ge \frac{n}{(n+1)M} -\frac{n-1}{n+1} \,.\]


Solution:

    1. \(\,\) \begin{align*} && 0 &\leq f(t) &\leq M \\ \Rightarrow && \int_0^x 0 \d t &\leq \int_0^x f(t) \d t & \leq \int_0^x M \d x \\ \Rightarrow && 0 &\leq F(x) &\leq Mx \end{align*}
    2. \(\,\) \begin{align*} && \int_0^1 2g(x)F(x)f(x) \d x &= \left [ g(x) F(x)^2 \right] - \int_0^1 g'(x) \left ( F(x)\right)^2 \d x \\ &&&= g(1) - \int_0^1 g'(x) \left ( F(x)\right)^2 \d x \end{align*}
    1. \(\,\) \begin{align*} && 1 &= \int_0^1 kF(y)f(y) \d y \\ &&&= k\left [ \frac12 F(y)^2\right]_0^1 \\ &&&= \frac{k}{2} \\ \Rightarrow && k &= 2 \end{align*}
    2. \(\,\) \begin{align*} \E[Y^n] &= \int_0^1 y^n 2F(y)f(y) \d y \\ &\geq \int_0^1 y^n 2My f(y) \d y \\ &= 2M\int_0^1 y^{n+1} f(y) \d y \\ &= 2M \E[X^{n+1}] = 2M\mu_{n+1} \\ \\ \E[Y^n] &= \int_0^1 y^n 2F(y)f(y) \d y \\ &= 1 - \int_0^1 ny^{n-1} F(y)^2 \d y \\ &\geq 1 - \int_0^1 ny^{n-1}My F(y) \d y \\ &= 1 - M\int_0^1 ny^n F(y) \d y \\ &= 1 - M[\frac{n}{n+1}y^{n+1} F(y)]_0^1 + M\int_0^1\frac{n}{n+1} y^{n+1} f(y) \d y \\ &= 1 - \frac{nM}{n+1} + \frac{nM}{n+1} \mu_{n+1} \end{align*}
    3. Since \(\E[Y^{n-1}] \geq 0\) we must have \begin{align*} && 2M\mu_n \geq 1 + \frac{(n-1)M}{n}\mu_n - \frac{(n-1)M}{n} \\ \Rightarrow && \mu_n \left (2M + \frac{(n-1)M}{n} \right) \geq 1 - \frac{(n-1)M}{n} \\ \Rightarrow && \mu_n \frac{3Mn-M}{n} & \geq \frac{n-(n-1)M}{n} \\ \Rightarrow && \mu_n & \geq \frac{n-(n-1)M}{3Mn-M} \end{align*}

2011 Paper 2 Q13
D: 1600.0 B: 1500.0

What property of a distribution is measured by its skewness?

  1. One measure of skewness, \(\gamma\), is given by \[ \displaystyle \gamma= \frac{ \E\big((X-\mu)^3\big)}{\sigma^3}\,, \] where \(\mu\) and \(\sigma^2\) are the mean and variance of the random variable \(X\). Show that \[ \gamma = \frac{ \E(X^3) -3\mu \sigma^2 - \mu^3}{\sigma^3}\,. \] The continuous random variable \(X\) has probability density function \(\f\) where \[ \f(x) = \begin{cases} 2x & \text{for } 0\le x\le 1\,, \\[2mm] 0 & \text{otherwise}\,. \end{cases} \] Show that for this distribution \(\gamma= -\dfrac{2\sqrt2}{5}\).
  2. The decile skewness, \(D\), of a distribution is defined by \[D= \frac { {\rm F}^{-1}(\frac9{10}) - 2{\rm F} ^{-1}(\frac12) + {\rm F}^{-1} (\frac1{10}) } {{\rm F}^{-1}(\frac9{10}) - {\rm F} ^{-1} (\frac1{10})}\,, \] where \({\rm F}^{-1}\) is the inverse of the cumulative distribution function. Show that, for the above distribution, \( D= 2 -\sqrt5\,.\) The Pearson skewness, \(P\), of a distribution is defined by \[ P = \frac{3(\mu-M)}{\sigma} \,,\] where \(M\) is the median. Find \(P\) for the above distribution and show that \(D > P > \gamma\,\).


Solution: Skewness is a measure of the symmetry (specifically the lack-thereof) in the distribution. How much mass is there on one side rather than another.

  1. \(\,\) \begin{align*} && \gamma &= \frac{\E \left [ (X - \mu)^3 \right ]}{\sigma^3} \\ &&&= \frac{\E \left [ X^3 - 3\mu X^2 + 3\mu^2 X - \mu^3 \right ]}{\sigma^3} \\ &&&= \frac{\E [ X^3 ]- 3\mu \E[X^2] + 3\mu^2 \E[X] - \mu^3 }{\sigma^3} \\ &&&= \frac{\E [ X^3 ]- 3\mu (\mu^2 + \sigma^2) + 3\mu^2\cdot \mu- \mu^3 }{\sigma^3} \\ &&&= \frac{\E [ X^3 ]- 3\mu \sigma^2 - \mu^3 }{\sigma^3} \\ \end{align*} \begin{align*} && f(x) &= \begin{cases} 2x & \text{for } 0\le x\le 1\,, \\[2mm] 0 & \text{otherwise}\,. \end{cases} \\ && \E[X] &= \int_0^1 2x^2 \d x \\ &&&= \frac23 \\ && \E[X^2] &= \int_0^1 2x^3 \d x \\ &&&= \frac12 \\ && \E[X^3] &= \int_0^1 2x^4 \d x \\ &&&= \frac25 \\ \\ && \mu &= \frac23 \\ && \sigma^2 &= \frac12 - \frac49 = \frac{1}{18} \\ && \gamma &= \frac{\frac25 - 3 \cdot \frac23 \cdot \frac1{18} - \frac8{27}}{\frac{1}{54\sqrt2}} \\ &&&= -\frac{2\sqrt2}{5} \end{align*}
  2. First note that \(\displaystyle F(x) = \int_0^x 2t \d t = x^2\) for \(x \in [0,1]\). In particular, \(F^{-1}(x) = \sqrt{x}\), so \begin{align*} && D &= \frac { {\rm F}^{-1}(\frac9{10}) - 2{\rm F} ^{-1}(\frac12) + {\rm F}^{-1} (\frac1{10}) } {{\rm F}^{-1}(\frac9{10}) - {\rm F} ^{-1} (\frac1{10})} \\ &&&= \frac{\sqrt{\frac9{10}} - 2 \sqrt{\frac5{10}} + \sqrt{\frac1{10}}}{\sqrt{\frac9{10}}-\sqrt{\frac1{10}}} \\ &&&= \frac{3-2\sqrt5+1}{3 - 1} \\ &&&= \frac{4-2\sqrt5}{2} = 2-\sqrt5 \end{align*} \begin{align*} && P &= \frac{3(\mu - M)}{\sigma} \\ &&&= \frac{3(\frac23 - \sqrt{\frac12})}{\frac{1}{3\sqrt2}} \\ &&&= 6 \sqrt2 - 9 \end{align*} First we compare \(P\) and \(D\), \(6\sqrt2-9\) and \(2-\sqrt5\) \begin{align*} && D & > P \\ \Leftrightarrow && 2-\sqrt5 &> 6\sqrt2 - 9 \\ \Leftrightarrow && 11 -6\sqrt2 &> \sqrt 5 \\ \Leftrightarrow && (121 + 72 - 132\sqrt2) & > 5 \\ \Leftrightarrow && 188 & > 132\sqrt2 \\ \Leftrightarrow && 47 & > 33 \sqrt 2\\ \Leftrightarrow && 2209 & > 2178 \end{align*} also \begin{align*} && P &> \gamma \\ \Leftrightarrow && 6\sqrt2 - 9 &> -\frac{2\sqrt2}{5} \\ \Leftrightarrow && 30\sqrt2 - 45 & > -2\sqrt2 \\ \Leftrightarrow && 32 \sqrt 2 &> 45 \\ \Leftrightarrow && 2048 &> 2025 \end{align*}

2011 Paper 3 Q12
D: 1700.0 B: 1516.0

The random variable \(N\) takes positive integer values and has pgf (probability generating function) \(\G(t)\). The random variables \(X_i\), where \(i=1\), \(2\), \(3\), \(\ldots,\) are independently and identically distributed, each with pgf \({H}(t)\). The random variables \(X_i\) are also independent of \(N\). The random variable \(Y\) is defined by \[ Y= \sum_{i=1}^N X_i \;. \] Given that the pgf of \(Y\) is \(\G(H(t))\), show that \[ \E(Y) = \E(N)\E(X_i) \text{ and } \var(Y) = \var(N)\big(\E(X_i)\big)^2 + \E(N) \var(X_i) \,.\] A fair coin is tossed until a head occurs. The total number of tosses is \(N\). The coin is then tossed a further \(N\) times and the total number of heads in these \(N\) tosses is \(Y\). Find in this particular case the pgf of \(Y\), \(\E(Y)\), \(\var(Y)\) and \(\P(Y=r)\).


Solution: Recall that for a random variable \(Z\) with pgf \(F(t)\) we have \(F(1) = 1\), \(\E[Z] = F'(1)\) and \(\E[Z^2] = F''(1) +F'(1)\) so \begin{align*} && \E[Y] &= G'(H(1))H'(1) \\ &&&= G'(1)H'(1) \\ &&&= \E[N]\E[X_i] \\ \\ && \E[Y^2] &= G''(H(1))(H'(1))^2+G'(H(1))H''(1) + G'(H(1))H'(1) \\ &&&= G''(1)(H'(1))^2+G'(1)H''(1) + G'(1)H'(1) \\ &&&= (\E[N^2]-\E[N])(\E[X_i])^2 + \E[N](\E[X_i^2]-\E[X_i]) + \E[N]\E[X_i] \\ &&&= (\E[N^2]-\E[N])(\E[X_i])^2 + \E[N]\E[X_i^2] \\ && \var[Y] &= (\E[N^2]-\E[N])(\E[X_i])^2 + \E[N]\E[X_i^2] - (\E[N])^2(\E[X_i])^2\\ &&&= (\var[N]+(\E[N])^2-\E[N])(\E[X_i])^2 + \E[N](\var[X_i]+\E[X_i]^2) - (\E[N])^2(\E[X_i])^2\\ &&&= \var[N](\E[X_i])^2 + \E[N]\var[X_i] \end{align*} Notice that \(N \sim Geo(\tfrac12)\) and \(Y = \sum_{i=1}^N X_i\) where \(X_i\) are Bernoulli. We have that \(G(t) = \frac{\frac12}{1-\frac12z}\) and \(H(t) = \frac12+\frac12p\) so the pgf of \(Y\) is \(G(H(t) = \frac{\frac12}{1 - \frac14-\frac14p} = \frac{2}{3-p}\). \begin{align*} && \E[X_i] &= \frac12\\ && \var[X_i] &= \frac14 \\ && \E[N] &= 2 \\ && \var[N] &= 2 \\ \\ && \E[Y] &= 2 \cdot \frac12 = 1 \\ && \var[Y] &= 2 \cdot \frac14 + 2 \frac14 = 1 \\ && \mathbb{P}(Y=r) &= \tfrac23 \left ( \tfrac13 \right)^r \end{align*}

2010 Paper 1 Q12
D: 1500.0 B: 1508.1

A discrete random variable \(X\) takes only positive integer values. Define \(\E(X)\) for this case, and show that \[\E(X) =\sum^{\infty}_{n=1}\P\left(X\ge n \right).\] I am collecting toy penguins from cereal boxes. Each box contains either one daddy penguin or one mummy penguin. The probability that a given box contains a daddy penguin is \(p\) and the probability that a given box contains a mummy penguin is \(q\), where \(p\ne0\), \(q\ne0\) and \(p+q=1\,\). Let \(X\) be the number of boxes that I need to open to get at least one of each kind of penguin. Show that \(\P(X\ge 4)= p^{3}+q^{3}\), and that \[ \E(X)=\frac{1}{pq}-1.\, \] Hence show that \(\E(X)\ge 3\,\).


Solution: \[ \E[X] := \sum_{n=1}^{\infty} n \mathbb{P}(X=n) \] \begin{align*} && \sum^{\infty}_{n=1}\mathbb{P}\left(X\ge n \right) &= \sum^{\infty}_{n=1}\sum_{k=n}^\infty \mathbb{P}(X=k) \\ &&&= \sum_{k=1}^\infty k \cdot \mathbb{P}(X=k) \\ &&&= \E[X] \end{align*} \begin{align*} &&\mathbb{P}(X \geq 4) &= \mathbb{P}(\text{first 3 are daddies}) +\mathbb{P}(\text{first 3 are mummies}) \\ &&&= p^3 + q^3 \\ \Rightarrow && \E[X] &= \sum_{n=1}^{\infty} \mathbb{P}\left(X\ge n \right) \\ &&&= 1+\sum_{n=2}^{\infty} \left ( p^{n-1} + q^{n-1}\right) \\ &&&= 1+\frac{p}{1-p} + \frac{q}{1-q} \\ &&&= 1+\frac{p}q + \frac{q}p \\ &&&= 1+\frac{p^2+q^2}{pq} \\ &&&= 1+\frac{(p+q)^2-2pq}{pq} \\ &&&= \frac{1}{pq} -1 \\ &&& \underbrace{\geq}_{AM-GM} \frac{1}{4}-1 = 3 \end{align*}

2010 Paper 2 Q12
D: 1600.0 B: 1486.8

The continuous random variable \(X\) has probability density function \(\f(x)\), where \[ \f(x) = \begin{cases} a & \text {for } 0\le x < k \\ b & \text{for } k \le x \le 1\\ 0 & \text{otherwise}, \end{cases} \] where \(a > b > 0\) and \(0 < k < 1\). Show that \(a > 1\) and \(b < 1\).

  1. Show that \[ \E(X) = \frac{1-2b+ab}{2(a-b)}\,. \]
  2. Show that the median, \(M\), of \(X\) is given by \(\displaystyle M=\frac 1 {2a}\) if \(a+b\ge 2ab\) and obtain an expression for the median if \(a+b\le 2ab\).
  3. Show that \(M < \E(X)\,\).


Solution: \begin{align*} && 1 &= \int_0^1 f(x) \d x \\ &&&= ak + b(1-k) \\ &&&= b + (a-b)k \\ \Rightarrow && k &= \frac{1-b}{a-b} \\ \Rightarrow && b & < 1 \tag{\(0 < k, \,a > b\)} \\ && k &> 1 \\ \Rightarrow && a-b & > 1-b \\ \Rightarrow && a > 1 \end{align*}

  1. \(\,\) \begin{align*} && \E[X] &= \int_0^1 x \cdot f(x) \d x \\ &&&= \int_0^k ax \d x + \int_k^1 b x \d x \\ &&&= a \frac{k^2}{2} + b \frac{1-k^2}{2} \\ &&&= \frac12b + (a-b) \frac{(1-b)^2}{2(a-b)^2} \\ &&&= \frac{(1-b)^2+b(a-b)}{2(a-b)} \\ &&&= \frac{1-2b+ab}{2(a-b)} \end{align*}
  2. \(\,\) The median \(M\) satisfies \[\frac12 = \int_0^M f(x) \d x \] If \(ka = \frac{a-ab}{a-b} \leq \frac12 \Leftrightarrow 2a-2ab \leq a-b \Leftrightarrow a+b \leq 2ab\) then \(M > k\) otherwise \(M < k\). In the latter case: \begin{align*} && \frac12 &= Ma \\ \Rightarrow && M &= \frac{1}{2M} \end{align*} In the former case \begin{align*} && \frac12 &= ka + (M-k)b \\ &&&= k(a-b) + Mb \\ &&&= 1-b + M b \\ \Rightarrow && M &= 1-\frac1{2b} \end{align*}

2010 Paper 3 Q13
D: 1700.0 B: 1516.0

In this question, \({\rm Corr}(U,V)\) denotes the product moment correlation coefficient between the random variables \(U\) and \(V\), defined by \[ \mathrm{Corr}(U,V) \equiv \frac{\mathrm{Cov}(U,V)}{\sqrt{\var(U)\var(V)}}\,. \] The independent random variables \(Z_1\), \(Z_2\) and \(Z_3\) each have expectation 0 and variance 1. What is the value of \(\mathrm{Corr} (Z_1,Z_2)\)? Let \(Y_1 = Z_1\) and let \[ Y_2 = \rho _{12} Z_1 + (1 - {\rho_{12}^2})^{ \frac12} Z_ 2\,, \] where \(\rho_{12}\) is a given constant with $-1<\rho _{12}<1$. Find \(\E(Y_2)\), \(\var(Y_2)\) and \(\mathrm{Corr}(Y_1, Y_2)\). Now let \(Y_3 = aZ_1 + bZ_2 + cZ_3\), where \(a\), \(b\) and \(c\) are real constants and \(c\ge0\). Given that \(\E(Y_3) = 0\), \(\var(Y_3) = 1\), \( \mathrm{Corr}(Y_1, Y_3) =\rho^{{2}}_{13} \) and \( \mathrm{Corr}(Y_2, Y_3)= \rho^{{2}} _{23}\), express \(a\), \(b\) and \(c\) in terms of \(\rho^{2} _{23}\), \(\rho^{2}_{13}\) and \(\rho^{2} _{12}\). Given constants \(\mu_i\) and \(\sigma_i\), for \(i=1\), \(2\) and \(3\), give expressions in terms of the \(Y_i\) for random variables \(X_i\) such that \(\E(X_i) = \mu_i\), \(\var(X_i) = \sigma_ i^2\) and \(\mathrm{Corr}(X_i,X_j) = \rho_{ij}\).


Solution: \begin{align*} \mathrm{Corr} (Z_1,Z_2) &= \frac{\mathrm{Cov}(Z_1,Z_2)}{\sqrt{\var(Z_1)\var(Z_2)}} \\ &= \frac{\mathbb{E}(Z_1 Z_2)}{\sqrt{1 \cdot 1}} \\ &= \frac{\mathbb{E}(Z_1)\mathbb{E}(Z_2)}{\sqrt{1 \cdot 1}} \\ &= \frac{0}{1} \\ &= 0 \end{align*} \begin{align*} && \mathbb{E}(Y_2) &= \mathbb{E}(\rho_{12} Z_1 + (1 - {\rho_{12}^2})^{ \frac12} Z_ 2) \\ &&&= \mathbb{E}(\rho_{12} Z_1) + \mathbb{E}( (1 - {\rho_{12}^2})^{ \frac12} Z_ 2) \\ &&&= \rho_{12}\mathbb{E}( Z_1) + (1 - {\rho_{12}^2})^{ \frac12}\mathbb{E}( Z_ 2) \\ &&&= 0\\ \\ && \textrm{Var}(Y_2) &= \textrm{Var}(\rho _{12} Z_1 + (1 - {\rho_{12}^2})^{ \frac12} Z_ 2) \\ &&&= \textrm{Var}(\rho_{12} Z_1)+\textrm{Cov}(\rho_{12} Z_1,(1 - {\rho_{12}^2})^{ \frac12} Z_ 2 ) + \textrm{Var}((1 - {\rho_{12}^2})^{ \frac12} Z_ 2) \\ &&&= \rho_{12}^2\textrm{Var}( Z_1)+\rho_{12} (1 - {\rho_{12}^2})^{ \frac12} \textrm{Cov}(Z_1, Z_ 2 ) + (1 - {\rho_{12}^2})\textrm{Var}(Z_ 2) \\ &&&= \rho_{12}^2 + (1-\rho_{12}^2) = 1 \\ \\ && \textrm{Cov}(Y_1, Y_2) &= \mathbb{E}((Y_1-0)(Y_2-0)) \\ &&&= \mathbb{E}(Z_1 \cdot (\rho _{12} Z_1 + (1 - {\rho_{12}^2})^{ \frac12} Z_ 2)) \\ &&&= \rho_{12} \mathbb{E}(Z_1^2) + (1-\rho_{12}^2)^{\frac12}\mathbb{E}(Z_1, Z_2) \\ &&&= \rho_{12} \\ \Rightarrow && \textrm{Corr}(Y_1, Y_2) &= \frac{\textrm{Cov}(Y_1, Y_2)}{\sqrt{\textrm{Var}(Y_1)\textrm{Var}(Y_2)}} \\ &&&= \frac{\rho_{12}}{1 \cdot 1} = \rho_{12} \end{align*} Suppose \(Y_3 =aZ_1 +bZ_2+cZ_3\) with \(\mathbb{E}(Y_3) = 0\) (must be true), \(\textrm{Var}(Y_3) = 1 = a^2+b^2+c^2\) and \(\textrm{Corr}(Y_1, Y_3) = \rho_{13}, \textrm{Corr}(Y_2, Y_3) = \rho_{23}\). \begin{align*} && \textrm{Corr}(Y_1,Y_3) &= \textrm{Cov}(Y_1, Y_3) \\ &&&= \textrm{Cov}(Z_1, aZ_1 +bZ_2+cZ_3) \\ &&&= a \\ \Rightarrow && a &= \rho_{13} \\ \\ && \textrm{Corr}(Y_2,Y_3) &= \textrm{Cov}(Y_2, Y_3) \\ &&&= \textrm{Cov}(\rho_{12}Z_1+(1-\rho_{12}^2)^\frac12Z_2, \rho_{13}Z_1 +bZ_2+cZ_3) \\ &&&= \rho_{12}\rho_{13}+(1-\rho_{12}^2)^\frac12b \\ \Rightarrow && \rho_{23} &= \rho_{12}\rho_{13}+(1-\rho_{12}^2)^\frac12b \\ \Rightarrow && b &= \frac{\rho_{23}-\rho_{12}\rho_{13}}{(1-\rho_{12}^2)^\frac12} \\ && c &= \sqrt{1-\rho_{13}^2-\frac{(\rho_{23}-\rho_{12}\rho_{13})^2}{(1-\rho_{12}^2)}} \end{align*} Finally, let \(X_i = \mu_i + \sigma_i Y_i\)

2009 Paper 1 Q13
D: 1500.0 B: 1504.1

I seat \(n\) boys and \(3\) girls in a line at random, so that each order of the \(n+3\) children is as likely to occur as any other. Let \(K\) be the maximum number of consecutive girls in the line so, for example, \(K=1\) if there is at least one boy between each pair of girls.

  1. Find \(\P(K=3)\).
  2. Show that \[\P(K=1)= \frac{n(n-1)}{(n+2)(n+3)}\,. \]
  3. Find \(\E(K)\).


Solution:

  1. If all the girls are say together there are \(n+1\) ways to place the block of 3 girls. There are \(\binom{n+3}{3}\) ways to choose where to place the girls in total, therefore: \begin{align*} && \mathbb{P}(K =3) &= \frac{n+1}{\binom{n+3}3} \\ &&&= \frac{6(n+1)}{(n+3)(n+2)(n+1)} \\ &&&= \frac{6}{(n+3)(n+2)} \end{align*}
  2. If \(K= 1\) then all of the girls are separated. We can place three girls and two boys separating them, then we are allocating \(N-2\) boys to \(4\) gaps, ie \(\binom{N-2+3}{3} = \binom{N+1}{3}\). \begin{align*} && \mathbb{P}(K=3) &= \frac{\binom{n+1}{3}}{\binom{n+3}{3}} \\ &&&= \frac{(n+1)n(n-1)}{(n+3)(n+2)(n+1)} \\ &&&= \frac{n(n-1)}{(n+3)(n+2)} \end{align*}
  3. \(\,\) \begin{align*} \mathbb{E}(K) &= \sum_{k=1}^3 k \mathbb{P}(K=k) \\ &= \frac{6}{(n+3)(n+2)} + 2 \left (1 - \frac{6}{(n+3)(n+2)} - \frac{n(n-1)}{(n+3)(n+2)} \right) + 3\frac{n(n-1)}{(n+3)(n+2)} \\ &= 2+\frac{6-12+n(n-1)}{(n+3)(n+2)} \\ &= 2 + \frac{n^2-n-6}{(n+2)(n+3)}\\ &= 2 + \frac{(n-3)(n+2)}{(n+2)(n+3)} \\ &= 2 + \frac{n-3}{n+3} \\ &= \frac{2n}{n+3} \end{align*}

2009 Paper 2 Q12
D: 1600.0 B: 1496.6

A continuous random variable \(X\) has probability density function given by \[ \f(x) = \begin{cases} 0 & \mbox{for } x<0 \\ k\e^{-2 x^2} & \mbox{for } 0\le x< \infty \;,\\ \end{cases} \] where \(k\) is a constant.

  1. Sketch the graph of \(\f(x)\).
  2. Find the value of \(k\).
  3. Determine \(\E(X)\) and \(\var(X)\).
  4. Use statistical tables to find, to three significant figures, the median value of \(X\).


Solution:

  1. \par
    TikZ diagram
  2. Let \(Y \sim N(0,\frac14)\), then: \begin{align*} &&\int_0^\infty \frac{1}{\sqrt{2 \pi \cdot \frac14}} e^{-2x^2} \, dx &= \frac12\\ \Rightarrow && \int_0^\infty e^{-2x^2} &= \frac{\sqrt{\pi}}{2 \sqrt{2}} \\ \Rightarrow && k &= \boxed{\frac{2\sqrt{2}}{\sqrt{\pi}}} \end{align*}
  3. \begin{align*} \mathbb{E}[X] &= \int_0^\infty x f(x) \, dx \\ &= \frac{2\sqrt{2}}{\sqrt{\pi}}\int_0^\infty x e^{-2x^2}\, dx \\ &= \frac{2\sqrt{2}}{\sqrt{\pi}} \left [-\frac{1}{4}e^{-2x^2} \right]_0^\infty \\ &= \frac{1}{\sqrt{2\pi}} \\ \end{align*} In order to calculate \(\mathbb{E}(X^2)\) it is useful to consider the related computation \(\mathbb{E}(Y^2)\). In fact, by symmetry, these will be the same values. Therefore \(\mathbb{E}(X^2) = \mathbb{E}(Y^2) = \mathrm{Var}(Y) = \frac{1}{4}\) (since \(\mathbb{E}(Y) = 0\)). Therefore \(\mathrm{Var}(Y) = \mathbb{E}(Y^2) - \mathbb{E}(Y)^2 = \frac14 - \frac{1}{2\pi}\)
  4. \begin{align*} && \mathbb{P}(X < x) &= \frac12 \\ \Leftrightarrow && 2\mathbb{P}(0 \leq Y < x) &= \frac12 \\ \Leftrightarrow && 2\l \mathbb{P}(Y < x) - \frac12 \r &= \frac12 \\ \Leftrightarrow && \mathbb{P}(Y < x)&= \frac34 \\ \Leftrightarrow && \mathbb{P}(\frac{Y-0}{1/2} < \frac{x}{1/2})&= \frac34 \\ \Leftrightarrow && \mathbb{P}(Z < \frac{x}{1/2})&= \frac34 \\ \Leftrightarrow && \Phi(2x)&= \frac34 \\ \Leftrightarrow && 2x &= 0.6744895\cdots \\ \Leftrightarrow && x &= 0.3372\cdots \\ \Leftrightarrow && &= 0.337 \, (3 \text{sf}) \\ \end{align*}

2008 Paper 1 Q12
D: 1516.0 B: 1484.0

In this question, you may use without proof the results: \[ \sum_{r=1}^n r = \tfrac12 n(n+1) \qquad\text{and}\qquad \sum_{r=1}^n r^2 = \tfrac1 6 n(n+1)(2n+1)\,. \] The independent random variables \(X_1\) and \(X_2\) each take values \(1\), \(2\), \(\ldots\), \(N\), each value being equally likely. The random variable \(X\) is defined by \[ X= \begin{cases} X_1 & \text { if } X_1\ge X_2\\ X_2 & \text { if } X_2\ge X_1\;. \end{cases} \]

  1. Show that \(\P(X=r) = \dfrac{2r-1}{N^2}\,\) for \(r=1\), \(2\), \(\ldots\), \(N\).
  2. Find an expression for the expectation, \(\mu\), of \(X\) and show that \(\mu=67.165\) in the case \(N=100\).
  3. The median, \(m\), of \(X\) is defined to be the integer such that \(\P(X\ge m) \ge \frac 12\) and \(\P(X\le m)\ge \frac12\). Find an expression for \(m\) in terms of \(N\) and give an explicit value for \(m\) in the case \(N=100\).
  4. Show that when \(N\) is very large, \[ \frac \mu m \approx \frac {2\sqrt2}3\,. \]


Solution: \begin{align*} \P(X = r) &= \P(X_1 = r, X_2 \leq r) + \P(X_2 = r, X_1 < r) \\ &= \P(X_1 = r) \P(X_2 \leq r) + \P(X_2 = r)\P( X_1 < r) \\ &= \frac{1}{N} \frac{r}{N} + \frac{1}{N} \frac{r-1}{N} \\ &= \frac{2r-1}{N^2} \end{align*} \begin{align*} \E(X) &= \sum_{r=1}^N r \P(X = r) \\ &= \sum_{r=1}^N \frac{2r^2 - r}{N^2} \\ &= \frac{1}{N^2} \l \frac{N(N+1)(2N+1)}{3} - \frac{N(N+1)}{2} \r \\ &= \frac{N+1}{N} \l \frac{4N-1}{6} \r \end{align*} When \(N = 100\), this is equal to \(\frac{101 \cdot 399}{6 \cdot 100} = \frac{101 \cdot 133}{200} = 67.165\) \begin{align*} &&\frac12 &\leq \P(X \leq m) \\ &&&=\sum_{r=1}^m \P(X=r) \\ &&&=\sum_{r=1}^m \frac{2r-1}{N^2} \\ &&&= \frac{1}{N^2} \l m(m+1) - m \r \\ &&&= \frac{m^2}{N^2} \\ \Rightarrow && m^2 &\geq \frac{N^2}{2} \\ \Rightarrow && m &\geq \frac{N}{\sqrt{2}} \\ \Rightarrow && m &= \left \lceil \frac{N}{\sqrt{2}} \right \rceil \end{align*} When \(N = 100\), \(100/\sqrt{2} = \sqrt{2}50\). \(\sqrt{2} > 1.4 \Rightarrow 50\sqrt{2} > 70\) \(\sqrt{2} < 1.42 \Rightarrow 50 \sqrt{2} < 71\), therefore \(\displaystyle \left \lceil \frac{100}{\sqrt{2}} \right \rceil = 71\) \begin{align*} \lim_{N \to \infty} \frac{\frac{(N+1)(4N-1)}{6N}}{ \left \lceil\frac{N}{\sqrt{2}} \right \rceil} &= \lim_{N \to \infty} \frac{\sqrt{2}}{3}\l \frac{4N^2 +3N - 1}{2N^2} \r \tag{since the floor will be irrelevant}\\ &= \lim_{N \to \infty} \frac{\sqrt{2}}{3}\l 2 + \frac{3}{2N} - \frac{1}{N^2} \r \\ &= \lim_{N \to \infty} \frac{2\sqrt{2}}{3} \end{align*}