| |
[Date Prev][Date Next][Thread Prev][Thread Next][Date index][Thread index]
Re: st: types of standard error
On May 23, 2006, at 9:03 AM, [email protected] wrote:
For example, in glm, we have the option of using:
OIM, EIM, OPG, HAC, jacknife, one-stepped jacknife, unbiased
sandwich...
If you really want to learn more about the differences between these
estimators, that's great; be forewarned, however, you're not going to
get very far without delving into some of the math. The distinction
between the expected information (EIM) and observed information (OIM)
is an important one, and comes out of basic likelihood theory (any
basic book on statistical theory will discuss this). In some cases
(e.g., generalized linear models with canonical link), they are
equivalent. A classic paper comparing the two is "Assessing the
accuracy of the maximum likelihood estimator: Observed versus
expected Fisher information" (Efron and Hinkley, Biometrika (65)
1978). This paper argues that the observed information is in general
more appropriate, which (I presume) is why -glm- produces this by
default.
The robust (sandwich) estimator is an attempt to relax the model
assumptions on which the standard calculation is based. For a simple
introduction (with examples), see "Model Robust Confidence Intervals
Using Maximum Likelihood Estimators" (Royall, International
Statistical Review (54) 1986). Also, as I said before, see [U] 20.14
and the references cited therein.
Efron and Tibshirani's book on the bootstrap (the one I suggested
earlier) actually treats both the bootstrap and the jackknife in
detail, and discusses the relationship between them and several other
estimates of variance (e.g., expected and observed information,
sandwich, etc.). In this respect, it is one of the better (and more
accessible) treatments of the differences among the various estimators.
Finally, if you're primarily interested in this issue WRT -glm-, I
believe Hardin and Hilbe's book (http://www.stata.com/bookstore/
glmext.html) has a good discussion of the differences between the
various estimators available.
Keep in mind, more choice (and the ease with which the different
estimates may be obtained and compared) is a very good thing. You
can never go wrong by trying the different estimators to verify that
they give similar results (which they often do). In cases where the
results are substantially different, however, you need to be careful
picking one over the others. And as always, they justifications for
many of these estimators rely on asymptotic arguments, and therefore
you should always be careful when applying them to data from small
samples.
-- Phil
*
* For searches and help try:
* http://www.stata.com/support/faqs/res/findit.html
* http://www.stata.com/support/statalist/faq
* http://www.ats.ucla.edu/stat/stata/