Paper 1, Section II, H

Statistics
Part IB, 2015

Suppose X1,,XnX_{1}, \ldots, X_{n} are independent identically distributed random variables each with probability mass function P(Xi=xi)=p(xi;θ)\mathbb{P}\left(X_{i}=x_{i}\right)=p\left(x_{i} ; \theta\right), where θ\theta is an unknown parameter. State what is meant by a sufficient statistic for θ\theta. State the factorisation criterion for a sufficient statistic. State and prove the Rao-Blackwell theorem.

Suppose that X1,,XnX_{1}, \ldots, X_{n} are independent identically distributed random variables with

P(Xi=xi)=(mxi)θxi(1θ)mxi,xi=0,,m\mathbb{P}\left(X_{i}=x_{i}\right)=\left(\begin{array}{c} m \\ x_{i} \end{array}\right) \theta^{x_{i}}(1-\theta)^{m-x_{i}}, \quad x_{i}=0, \ldots, m

where mm is a known positive integer and θ\theta is unknown. Show that θ~=X1/m\tilde{\theta}=X_{1} / m is unbiased for θ\theta.

Show that T=i=1nXiT=\sum_{i=1}^{n} X_{i} is sufficient for θ\theta and use the Rao-Blackwell theorem to find another unbiased estimator θ^\hat{\theta} for θ\theta, giving details of your derivation. Calculate the variance of θ^\hat{\theta} and compare it to the variance of θ~\tilde{\theta}.

A statistician cannot remember the exact statement of the Rao-Blackwell theorem and calculates E(TX1)\mathbb{E}\left(T \mid X_{1}\right) in an attempt to find an estimator of θ\theta. Comment on the suitability or otherwise of this approach, giving your reasons.

[Hint: If aa and bb are positive integers then, for r=0,1,,a+b,(a+br)=r=0,1, \ldots, a+b,\left(\begin{array}{c}a+b \\ r\end{array}\right)= j=0r(aj)(brj).]\left.\sum_{j=0}^{r}\left(\begin{array}{c}a \\ j\end{array}\right)\left(\begin{array}{c}b \\ r-j\end{array}\right) .\right]