Problems

Filters
Clear Filters

2 problems found

2014 Paper 3 Q12
D: 1700.0 B: 1500.0

The random variable \(X\) has probability density function \(f(x)\) (which you may assume is differentiable) and cumulative distribution function \(F(x)\) where \(-\infty < x < \infty \). The random variable \(Y\) is defined by \(Y= \e^X\). You may assume throughout this question that \(X\) and \(Y\) have unique modes.

  1. Find the median value \(y_m\) of \(Y\) in terms of the median value \(x_m\) of \(X\).
  2. Show that the probability density function of \(Y\) is \(f(\ln y)/y\), and deduce that the mode \(\lambda\) of \(Y\) satisfies \(\f'(\ln \lambda) = \f(\ln \lambda)\).
  3. Suppose now that \(X \sim {\rm N} (\mu,\sigma^2)\), so that \[ f(x) = \frac{1}{\sigma \sqrt{2\pi}\,} \e^{-(x-\mu)^2/(2\sigma^2)} \,. \] Explain why \[\frac{1}{\sigma \sqrt{2\pi}\,} \int_{-\infty}^{\infty}\e^{-(x-\mu-\sigma^2)^2/(2\sigma^2)} \d x = 1 \] and hence show that \( \E(Y) = \e ^{\mu+\frac12\sigma^2}\).
  4. Show that, when \(X \sim {\rm N} (\mu,\sigma^2)\), \[ \lambda < y_m < \E(Y)\,. \]


Solution:

  1. \begin{align*} && \frac12 &= \mathbb{P}(X \leq x_m) \\ \Leftrightarrow && \frac12 &= \mathbb{P}(e^X \leq e^{x_m} = y_m) \end{align*} Therefore the median is \(y_m = e^{x_m}\)
  2. \begin{align*} && \mathbb{P}(Y \leq y) &= \mathbb{P}(e^X \leq y) \\ &&&= \mathbb{P}(X \leq \ln y) \\ &&&= F(\ln y) \\ \Rightarrow && f_Y(y) &= f(\ln y)/y \\ \\ && f'_Y(y) &= \frac{f'(\ln y) - f(\ln y)}{y^2} \end{align*} Therefore since the mode satisfies \(f'_Y = 0\) we must have \(f'(\ln \lambda ) = f(\ln \lambda)\)
  3. This is the integral of the pdf of \(N(\mu + \sigma^2, \sigma^2)\) and therefore is clearly \(1\). \begin{align*} && \E[Y] &= \int_{-\infty}^{\infty} e^x \cdot \frac{1}{\sqrt{2\pi \sigma^2}} e^{-(x-\mu)^2/(2\sigma^2)} \d x \\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (x - (x-\mu)^2/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp ((2x \sigma^2- (x-\mu)^2)/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu-\sigma^2)^2+2\mu \sigma^2-\sigma^4)/(2\sigma^2)) \d x\\ &&&= \frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu+\sigma^2)^2)/(2\sigma^2)+\mu +\frac12\sigma^2) \d x\\ &&&= \e^{\mu +\frac12\sigma^2}\frac{1}{\sqrt{2\pi \sigma^2}} \int_{-\infty}^{\infty} \exp (-(x-\mu-\sigma^2)^2)/(2\sigma^2)) \d x\\ &&&= \e^{\mu +\frac12\sigma^2} \end{align*}
  4. Notice that \(y_m = e^\mu < e^{\mu + \tfrac12 \sigma^2} = \E[Y]\), so it suffices to prove that \(\lambda < e^{\mu}\) Notice that \(f'(x) - f(x) = f(x)[-(x-\mu)/\sigma^2 - 1]\) and therefore \(\ln y - \mu = -\sigma^2\) so \(\lambda = e^{\mu - \sigma^2}\) which is clearly less than \(e^{\mu}\) as required.

1997 Paper 3 Q13
D: 1700.0 B: 1500.0

Let \(X\) and \(Y\) be independent standard normal random variables: the probability density function, \(\f\), of each is therefore given by \[ \f(x)=\left(2\pi\right)^{-\frac{1}{2}}\e^{-\frac{1}{2}x^{2}}. \]

  1. Find the moment generating function \(\mathrm{E}(\e^{\theta X})\) of \(X\).
  2. Find the moment generating function of \(aX+bY\) and hence obtain the condition on \(a\) and \(b\) which ensures that \(aX+bY\) has the same distribution as \(X\) and \(Y\).
  3. Let \(Z=\e^{\mu+\sigma X}\). Show that \[ \mathrm{E}(Z^{\theta})=\e^{\mu\theta+\frac{1}{2}\sigma^{2}\theta^{2}}, \] and hence find the expectation and variance of \(Z\).


Solution:

  1. \(\,\) \begin{align*} && \E[e^{\theta X}] &= \int_{-\infty}^{\infty} e^{\theta x} \frac{1}{\sqrt{2\pi}} e^{-\frac12 x^2 } \d x\\ &&&= \int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}} e^{-\frac12 x^2+\theta x} \d x\\ &&&= \int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}} e^{-\frac12 (x^2-2\theta x)} \d x\\ &&&= \int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}} e^{-\frac12 (x-\theta )^2+\frac12\theta^2 } \d x\\ &&&= e^{\frac12\theta^2 }\int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}} e^{-\frac12 (x-\theta )^2 } \d x\\ &&&=e^{\frac12\theta^2 } \end{align*}
  2. \begin{align*} && M_{aX+bY} (\theta) &= \mathbb{E}[e^{\theta (aX+bY)}] \\ &&&= e^{\frac12(a\theta)^2} \cdot e^{\frac12(b\theta)^2} \\ &&&= e^{\frac12(a^2+b^2)\theta^2} \end{align*} Therefore we need \(a^2+b^2 = 1\)
  3. \(\,\) \begin{align*} && \E[Z^\theta] &= \E[e^{\mu \theta + \sigma \theta X}] \\ &&&= e^{\mu \theta}e^{\frac12 \sigma^2 \theta^2} \\ &&&=e^{\mu \theta + \frac12 \sigma^2 \theta^2} \\ \end{align*} \begin{align*} \mathbb{E}(Z) &= \mathbb{E}[Z^1] \\ &= e^{\mu + \frac12 \sigma^2} \\ \var[Z] &= \E[Z^2] - \left ( \E[Z] \right)^2 \\ &= e^{2 \mu+ 2\sigma^2} - e^{2\mu + \sigma^2} \\ &= e^{2\mu+\sigma^2} \left (e^{\sigma^2}-1 \right) \end{align*} [NB: This is the lognormal distribution]