Suppose that X1 Xn are independent random variables with density or mass functio
ID: 3390045 • Letter: S
Question
Suppose that X1 Xn are independent random variables with density or mass function f(x, theta) and suppose that we estimate theta using the maximum likelihood estimator theta: we estimate its standard error using the observed Fisher information estimator se(theta)={ -nsigma i=1 l"(x, theta)^-1/2} where l'(x,theta),l"(x, theta) are the first two partial derivatives of In f(x, teta) with respect to theta Alternatively, we could use the jackknife to estimate the standard error of theta. if our model is correct then we would expect (hope) that the two estimates are similar. In order to investigate this, we need to be able to get a good approximation to the ''leave-one-out" estimators {theta-i,}. Show that theta-i, satisfies the equation n l'xi,theta =sigma j=1 l(xi, theta-iExplanation / Answer
Suppose we have samplevX=(x1,x2,....xn) and estimator ^=S(X), then the jacknife focuses on the samples that leave out one observation at a time is X(i)=(x1,x2,.....,xi-1,xi+1,......xn) is called jacknife samples. Then the jacknife estimate of se is defined as se^2jack=[(n-1)/n(summation of i=1 to n(^(i)-^(.))2)]. By Fisher -Neyman criterian L=(product of i=1 to n)f(xi,) and l'(x,)=0 and l"(x,)<0 then l'(Xi;^(-i))=(summation j=1 to n)li(Xj;^(-i)).
Related Questions
Hire Me For All Your Tutoring Needs
Integrity-first tutoring: clear explanations, guidance, and feedback.
Drop an Email at
drjack9650@gmail.com
drjack9650@gmail.com
Navigate
Integrity-first tutoring: explanations and feedback only — we do not complete graded work. Learn more.