+0

# Statistical Inference Information Identity

0
687
0

I'm in this course called Statistical Inference and I was hoping someone could help me out with this exercise.

The trouble I'm having with is in part 4, but I'll give the answers to 1-3 to make it easier for the reader.

This is the exercise; 1.

$$\begin{array}{lcl} \mbox{Let }L(\theta;y) \mbox{ be the likelihood}\\ \mbox{Let }l(\theta;y) \mbox{ be the log-likelihood }(l(\theta;y) = ln(L(\theta;y)))\\ \mbox{Let }u(\theta;y) \mbox{ be the score function}(u(\theta;y) = \frac{\partial l(\theta;y)}{\partial \theta})\\ \mbox{Then it can be shown that}\\ u(\theta;y)= \frac{n}{\theta}-\sum^n_{i=1}ln(y_i) \end{array}$$

2.

$$\begin{array}{lcl} \mbox{Let }\hat{\theta} \mbox{ be the MLE}\\ \mbox{Then it should hold that }u(\hat{\theta};y) = 0\\ \mbox{solving this gives}\\ \hat{\theta} = \frac{n}{\sum^n_{i=1}ln(y_i)} \end{array}$$

3.

$$\begin{array}{lcl} \mbox{The expected fisher information } I(\theta) \mbox{ is given as }\\ I(\theta) = -E[\frac{\partial u}{\theta} (\theta;y)]\\ \mbox{By computation;}\\ I(\theta) = \frac{n}{\theta^2} \end{array}$$

So, for exercise 4, I can find that I also have the answer, which uses the change of variable technique with x:= ln(y), and the expected fisher information associated to a single r.v. (i(theta)), but I don't have a clue what they're doing there.

Could anyone help me out? 