I have two groups (A and B) that each completed a different version of an experimental measure (measure A and measure B). Both groups also completed a standard, "gold standard", measure. For the purposes of my study, I am treating the "gold standard" measure as a participant's true score. If I look only at group means, there does not appear to be any difference between the measures. Both groups obtained very similar scores on their respective versions of the experimental measure, and both groups obtained similar means on the "gold standard" measure. However, for group A, I noticed that there was a lot more variance in their scores on measure A, than for group B in the variance in their scores for measure B. So, although measure A and measure B did about as well each other and obtained the same means as the gold standard, group A appears to be a lot more unreliable when considering the scores of individual participants. Basically, people who completed measure A tended to give more extreme scores, both overshooting and undershooting their true score (obtained by the gold standard measure), but somehow, they cancelled each other out and their group means was the same as group B's. This difference in variance is also noticeable when I look at correlation coefficients, as measure A has a lower correlation with the gold standard measure than measure B. I compared the correlation coefficients with a Fisher Z test, and the differences in correlation coefficient are statistically significant.
So, here's where I'm hung up:
I'm wondering if it would be okay to calculate the absolute value of the difference score between the gold standard and the experimental measure for each participant. I would then want to compare the means for this absolute values of the differences for each group, with a t-test or ANOVA. I think this would be more meaningful than simply comparing correlation coefficients, because it would help me quantify in meaningful units, on average, how much more each participant completing measure A missed their true score, and then make a statement about the significance of the difference between the groups. Is this Kosher? Is there a better procedure for me? I hope the information I provided wasn't too vague, but I am a first time poster and a little shy about asking questions on a public forum.
So, here's where I'm hung up:
I'm wondering if it would be okay to calculate the absolute value of the difference score between the gold standard and the experimental measure for each participant. I would then want to compare the means for this absolute values of the differences for each group, with a t-test or ANOVA. I think this would be more meaningful than simply comparing correlation coefficients, because it would help me quantify in meaningful units, on average, how much more each participant completing measure A missed their true score, and then make a statement about the significance of the difference between the groups. Is this Kosher? Is there a better procedure for me? I hope the information I provided wasn't too vague, but I am a first time poster and a little shy about asking questions on a public forum.