|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: st: Increasing variance of dependent variable, logit, inter-rater agreement
From |
Steven Samuels <[email protected]> |
To |
[email protected] |
Subject |
Re: st: Increasing variance of dependent variable, logit, inter-rater agreement |
Date |
Fri, 27 Feb 2009 14:51:07 -0500 |
You have given very little information about these technologies and
what they are measuring. I do not know much about this area, but have
a few thoughts and questions for you to answer. If you do, perhaps
others may be able to respond to your problem.
Please give more detail about what is being assessed. Is there a gold
standard, measured or latent, for what these technologies are trying
to agree upon?
* What is the first technology that measures characteristics and
arrives at a pass-fail? How does it make this decision? Was age one
of these characteristics?
* How was the cut point y2b arrived at?
* You say that the variability of y2a increases with age. Is the
level of y2a related to age?
* "Agreement" is usually defined in two ways: 1) agreement on the
marginal proportions of the outcomes; 2) agreement of individual
decisions. The first is traditionally assessed by McNemar's test;
the second by kappa. (See any book on categorical data).
Kappa corrects for chance agreement. The problem with measuring just
the percentage of agreement, as you propose, is that it may be quite
high by chance. Take two raters who independently make decisions with
random numbers. Assume that each scores a "fail" with probability
0.20. The probability that they agree will be 0.68, purely by chance.
Kappa is not easy to model, but you could do it if you modeled all
four cells of the 2x2 table of technology one vs technology two with
a multinomial logistic or probit, followed by -nlcom-. However, I
think that a jackknife or bootstrap confidence interval would be
better. If age (perhaps through -fracpoly-) is a predictor, this
might account for the age-related SD of the measurement.
-Steve
On Feb 26, 2009, at 8:30 PM, Supnithadnaporn, Anupit wrote:
Dear Stata List members,
I am trying to learn what factors might determine the inter-rater
agreement
between the two tests of the same subject using different
technologies. One technology measures the characterisitcs of a
subject and gives the result of
pass-fail (y1). The other technology gives the interval measurement
(y2a) and the pass-fail (y2b) can be obtained using certain threshold.
My simple idea is to use the logit regression of a dependent variable
called 'agree' which takes value 1 if both tests yield the same
results
(y1=y2b) otherwise 0.
agree = f(age, x1,x2,x3,...)
The problem is that as 'age' increases, I know that the variance of
y2a
also increases. This is due to the nature (deterioration) of the
subject
itself. Given that the quality of the instrument measuring y2a is
constant,
how should I take into account the increasing variance of y2a in my
model?
*
* For searches and help try:
* http://www.stata.com/help.cgi?search
* http://www.stata.com/support/statalist/faq
* http://www.ats.ucla.edu/stat/stata/