[R-sig-ME] Extracting Standard Errors of Uncorrelated Random Effects?

Andrew Robinson A.Robinson at ms.unimelb.edu.au
Thu Dec 15 07:07:12 CET 2011


On Wed, Dec 14, 2011 at 03:34:18PM -0600, Douglas Bates wrote:
> 
> There are differences of opinion on this.  I have spent the last 30+
> years in the Department of Statistics at the University of Wisconsin
> - Madison, a department that was founded by George Box.  To me it is
> natural to pursue parsimony in a model, which means that I delete
> terms that do not appear to be contributing significantly to the
> model fit.  Like the quote attributed to Einstein, "Make things as
> simple as possible, but no simpler."
> 
> In fact, in this particular case it is of interest to test the
> hypothesis of no correlation of the random effects because the
> experimenters want to know if they can predict the extent of sleep
> deprivation's effect on response time from the initial response time.
> (I.e., are those who have faster response times initially less
> affected by sleep deprivation?)
> 
> Others (feel free to chime in here, Ben) believe that "model building"
> by deleting apparently insignificant terms results in overfitting of
> the model and I don't dispute that.  In some ways it depends on what
> the objective in fitting the model is.  For the purposes of prediction
> I want to pay attention to the bias-variance trade-off and aim for a
> simple, adequate model.  For the purposes of establishing the
> significance of a fixed-effects term, preliminary simplification of
> the model may bias the effect of the model.

I think that it really does depend on the objective in fitting the
model and the provenance of the data.  If an experimental design has
been established and one wishes to perform some inference conditional
on this design, then it is essential that the design be reflected in
the model structure.  Hence, in my opinion, testing some random
effects for inclusion makes no sense, but others are fair game, and
which is which depends on the design.
 
Cheers

Andrew

> The part of the "stay with the initial model regardless" approach that
> I don't like is that I am not convinced that the initial model is
> necessarily a good model.
> 
> 
> > Many thanks again for your excellent work on lme4.
> > Derek
> > --
> > Derek Dunfield, PhD
> > Postdoctoral Fellow, MIT Intelligence Initiative
> > Sloan School of Management
> > MIT Center for Neuroeconomics, Prelec Lab
> > 77 Massachusetts Ave E62-585
> > Cambridge MA 02139
> >
> >
> >
> > On Wed, Dec 14, 2011 at 3:16 PM, Douglas Bates <bates at stat.wisc.edu> wrote:
> >> On Wed, Dec 14, 2011 at 1:05 PM, Derek Dunfield <dunfield at mit.edu> wrote:
> >>> Hello all:
> >>> I was hoping to extract SE for my random effects to get a better idea
> >>> of their precision for each grouping factor.
> >>>
> >>> Andrew Gelman has developed a nice package called "Arm" which does
> >>> this seamlessly for models with correlated random effects such as
> >>> Reaction ~ 1 + Days + (1 + Days | Subject)
> >>>
> >>> When I try to run this on an uncorrelated model such as
> >>> Reaction ~ 1 + Days + (1 | Subject) + (0 + Days | Subject)
> >>>
> >>> I get the error "Error in .local(object, ...) : Code not written yet"
> >>
> >> Unfortunately, it is complicated.  Strangely enough, the case of
> >> correlated random effects is easier than the case of uncorrelated
> >> effects.  The values that you want to get are the solutions to a
> >> system like L x = cbind(e_i, e_j) where the indices i and j are the
> >> random effects corresponding to the k'th level of the grouping factor
> >> and L is the sparse Cholesky factor.  For correlated random effects
> >> the i and j are going to be adjacent.  For uncorrelated random effects
> >> they may end up being adjacent after the fill-reducing permutation but
> >> that is not guaranteed.
> >>
> >> In this particular case things work out fine.  If I fit that model as
> >> fm1 then I can check that the random effects for each subject are
> >> adjacent by examining
> >>
> >> image(fm1 at L)
> >>
> >> where we can see that each of the pairs of blocks starting with 1:2
> >> are correlated within the block but orthogonal to the other blocks.
> >> So, setting
> >>
> >> L <- fm1 at L
> >> Id <- Diagonal(x = rep.int(1, ncol(L))   # Identity matrix of the same size as L
> >>
> >>  the conditional variance-covariance matrix of the orthogonal random
> >> effects for subject 1 is something like
> >>
> >> crossprod(solve(L, Id[,1:2], sys="L")) * lme4:::sigma(fm1)^2
> >>
> >> Before using such an expression I would check it against the results
> >> from the "Arm" package for the model with correlated random effects.
> >> It is entirely possible that I have left something out of the
> >> calculation.
> >>
> >> If you wonder why something that looks simple like this is not part of
> >> the lme4 package, it is because a lot of the simplicity comes from the
> >> fact that both random-effects terms have the same grouping factor.
> >> The general case is much messier.
> >>
> >>> I believe this may be due to some unfinished code in the lme4 package
> >>> for ranef, though I'm not sure.
> >>>
> >>> Does anyone have an idea how I could go about calculating these standard errors?
> >>>
> >>> Many thanks in advance for you help!
> >>> :Derek
> >>>
> >>> --
> >>> Derek Dunfield, PhD
> >>> Postdoctoral Fellow, MIT Intelligence Initiative
> >>> Sloan School of Management
> >>> MIT Center for Neuroeconomics, Prelec Lab
> >>> 77 Massachusetts Ave E62-585
> >>> Cambridge MA 02139
> >>>
> >>> _______________________________________________
> >>> R-sig-mixed-models at r-project.org mailing list
> >>> https://stat.ethz.ch/mailman/listinfo/r-sig-mixed-models
> 
> _______________________________________________
> R-sig-mixed-models at r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-sig-mixed-models

-- 
Andrew Robinson  
Deputy Director, ACERA 
Department of Mathematics and Statistics            Tel: +61-3-8344-6410
University of Melbourne, VIC 3010 Australia               (prefer email)
http://www.ms.unimelb.edu.au/~andrewpr              Fax: +61-3-8344-4599
http://www.acera.unimelb.edu.au/

Forest Analytics with R (Springer, 2011) 
http://www.ms.unimelb.edu.au/FAwR/
Introduction to Scientific Programming and Simulation using R (CRC, 2009): 
http://www.ms.unimelb.edu.au/spuRs/




More information about the R-sig-mixed-models mailing list