[R] Anova - adjusted or sequential sums of squares?
Douglas Bates
bates at stat.wisc.edu
Wed Apr 20 17:19:09 CEST 2005
Michael Watson (IAH-C) wrote:
> I guess the real problem is this:
>
> As I have a different number of observations in each of the groups, the
> results *change* depending on which order I specify the factors in the
> model. This unnerves me. With a completely balanced design, this
> doesn't happen - the results are the same no matter which order I
> specify the factors.
>
> It's this reason that I have been given for using the so-called type III
> adjusted sums of squares...
>
Yes, but how do you know that the test represented by the type III sums
of squares makes sense in the context of your data?
The point that Bill is trying to make is that hypothesis tests always
involve comparing the fits of two nested models to some data. If you
can't describe what the models being compared are, how can you interpret
the results of the test?
There are many concepts introduced in statistics that apply to certain,
specific cases but have managed to outgrow the original context so that
people think they have general application. The area of linear models
and the analysis of variance is overrun with such concepts. The list
includes "significance of main effects", "overall mean", "R^2",
"expected mean square", ... These concepts do not stand on their own -
they apply to specific models.
You are looking for *the answer* to the question "Is this main effect
significant?" What Bill is saying is that the question doesn't make
sense. You can ask "Does the model that incorporates this term and all
other first-order terms provide a significantly better fit than same
model without this one term?" That's a well-phrased question. You
could even ask the same question about a model with first-order and
higher-order terms versus the same model without this one term and you
can get an answer to that question. Whether or not that answer makes
sense depends on whether or not the model with all the terms except the
one being considered makes sense. In most cases it doesn't so why say
that you must get a p-value for a nonsensical test. That number does
*not* characterize a test of the "significance of the main effect".
How does R come in to this? Well, with R you can fit models of great
complexity to reasonably large data sets quickly and easily so, if you
can formulate the hypothesis of interest to you by comparing the fit of
an inner model to an outer model, then you simply fit them and compare
them, usually with anova(fm1, fm2). That's it. That's all that the
analysis of variance is about. The complicated formulas and convoluted
reasoning that we were all taught are there solely for the purpose of
trying to "simplify" the calculations for this comparison. They're
unnecessary. With a tool like R you simple fit model 1 then fit model 2
and compare the fits. The only kicker is that you have to be able to
describe your hypothesis in terms of the difference between two models.
With tools like R we have the potential to change statistics is viewed
as a discipline and especially the way that it is taught. Statistics is
not about formulas - statistics is about models. R allows you to think
about the models and not grubby details of the calculations.
More information about the R-help
mailing list