Notice: On April 23, 2014, Statalist moved from an email list to a forum, based at statalist.org.
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: st: RE: RE: two sample test under generalized Behrens-Fisher conditions
From
Steven Samuels <[email protected]>
To
[email protected]
Subject
Re: st: RE: RE: two sample test under generalized Behrens-Fisher conditions
Date
Tue, 14 Dec 2010 12:09:34 -0500
I don't think that highly of t-tests. To quote Hampel et al. (Robust
Statistics: The Approach Based on Influence Functions, Wiley, NY,
1986) p. 405:
"Many statisticians are proud of the so-called robustness of the t-
test and more generally of the test in fixed-effects models in the
analysis of variance. But this robustness is only a rather moderate
and limited robustness of level ("robustness of validity"); the power
("robustness of efficiencey") and hence also the length of confidence
intervals and the size of standard erros is very nonrobust.
Consequently, a significant result can be believed, but non-
significance may just be due to the inefficiency of least squares."
Perhaps the easiest alternative to teach would be one based on trimmed
means, which are not only easy to understand (as opposed to, say, M-
Estimators and robust regression), but, unlike the median, have an
easy standard error formula.
Steve
[email protected]
On Dec 14, 2010, at 10:16 AM, Nick Cox wrote:
I see the problem. I couldn't (wouldn't) fit -glm- in an introductory
course either.
In similar circumstances I usually assert that t tests work well even
if the assumptions are not well satisfied. This is an idea that goes
back at least to G.E.P. Box in Biometrika 1953:
Box, G.E.P. 1953. Non-normality and tests on variances. Biometrika 40:
318-35.
Nick
[email protected]
Airey, David C
I was looking for "stark cookbooky" solutions for a (too) short intro
course that will not address GLM. But transformations they will be
told about, and the last time I taught this course, your help file
about transformations was required reading. Thanks for that citation.
Looks like a good book.
Nick Cox
In this kind of territory, I would always
1. Check out what is said in Rupert G. Miller, Beyond ANOVA. See on
the CRC Press reissue
<
http://www.crcpress.com/utility_search/search_results.jsf?conversationId=250169
Your library may hold a copy of the Wiley original.
2. Be wary of the stark cookbooky alternative: data if normal, ranks
otherwise. What happened to the idea of transformations or link
functions? How do you decide when the data are approximately normal
any way?
Here is an example of a different approach. In the auto data, -mpg-
given -foreign- is neither normal nor heteroscedastic. But these are
secondary issues. Consider this set of results. In each -
family(normal)- is implied.
foreach v in "power 1" "power 0.5" "log" "power -0.5" "power -1" {
qui glm mpg foreign, link(`v')
mat b = e(b)
mat V = e(V)
di "`v'" "{col 20}" %3.2f b[1,1] / sqrt(V[1,1])
}
power 1 3.63
power 0.5 3.70
log 3.75
power -0.5 -3.78
power -1 -3.80
The change of sign of what -glm- calls the z statistic is an
expected side-effect of changing to inverse transformations. More
importantly, z changes only very slowly and the collective set of
results points to the idea that 1/mpg is a more appropriate scale
than mpg on which to test for differences. This of course matches
basic science.
Generalized linear models are nearly 40 years old as a family. When
are they going to receive the recognition they deserve?
Airey, David C
I was reading a little about what to do when you have both unequal
variance and non-normality. Neither the equal variance t-test nor
the Mann-Whitney U test are best when you want to interpret the
difference in means or medians.
I had found the Stata command -fprank-, but it turns out this
robust ranks test doesn't escape a symmetry assumption to interpret
the location difference.
I found that some recommend using Welch's t-test on the ranked data
(Zimmerman and Zumbo (1993) Rank transformations and the power of
the Student's t test and the Welch t' test for non-normal
populations with unequal variances. Canadian Journal of
Experimental Psychology 47:3, 523-539).
This appears easy and satisfying solution to teach with: always use
unequal variances t-test and use ranks if the data are also not
normal.
*
* For searches and help try:
* http://www.stata.com/help.cgi?search
* http://www.stata.com/support/statalist/faq
* http://www.ats.ucla.edu/stat/stata/
*
* For searches and help try:
* http://www.stata.com/help.cgi?search
* http://www.stata.com/support/statalist/faq
* http://www.ats.ucla.edu/stat/stata/