(a) Define the maximum likelihood estimator (MLE) and the Fisher information I(θ).
(b) Let Θ=R and assume there exist a continuous one-to-one function μ:R→R and a real-valued function h such that
Eθ[h(X)]=μ(θ)∀θ∈R
(i) For X1,…,Xn i.i.d. from the model for some θ0∈R, give the limit in almost sure sense of
μ^n=n1i=1∑nh(Xi)
Give a consistent estimator θ^n of θ0 in terms of μ^n.
(ii) Assume further that Eθ0[h(X)2]<∞ and that μ is continuously differentiable and strictly monotone. What is the limit in distribution of n(θ^n−θ0). Assume too that the statistical model satisfies the usual regularity assumptions. Do you necessarily expect Var(θ^n)⩾(nI(θ0))−1 for all n ? Why?
(iii) Propose an alternative estimator for θ0 with smaller bias than θ^n if Bn(θ0)=Eθ0[θ^n]−θ0=na+n2b+O(n31) for some a,b∈R with a=0.
(iv) Further to all the assumptions in iii), assume that the MLE for θ0 is of the form
θ^MLE=n1i=1∑nh(Xi)
What is the link between the Fisher information at θ0 and the variance of h(X) ? What does this mean in terms of the precision of the estimator and why?
[You may use results from the course, provided you state them clearly.]