[R-sig-ME] GLMER vs. MCMCglmm - difference in model results - Poisson model with offset term

Bernd Lenzner bernd@|enzner @end|ng |rom un|v|e@@c@@t
Thu May 23 16:50:23 CEST 2019


Dear everyone,

I have a question regarding the comparison of Generalized Linear Mixed 
Models (GLMM - using the lme4 - package) and Markov-Chain Monte-Carlo 
Generalized Mixed Models (MCMCglmm - using the MCMCglmm-package).

I am running mixed models using both packages. While the estimates 
provided by each approach are relatively comparable, the significance 
levels differ strongly. My main questions are the following:

1. How can the divergence in results be explained and why doe both 
methods provide different results even though the model structure is the 
same? (Note: diagnostic plots for both models look fine)

2. I am aware that there is strong debate about the validity of p-values 
within GLMMs. Is there similar concern regarding MCMCglmm?

3. What would be a sensible interpretation of both model outputs 
especially with respect to PRED2 and the interaction term.


Below I provide a short description and example output:

I want to test the effect of 3 continuous predictors (PRED2, PRED3, 
PRED4) on a proportional response (RESP). PRED2 and PRED3 enter the 
model as an interaction effect (PRED2*PRED3). The response is modeled 
using a Poisson error structure with a log-link function and the 
nominator of the ratio is therefore included as an offset predictor 
(offset(log(PRED1))).


MODEL1 -  GLMM using the lme4-package:


MOD1 <- glmer(RESP ~ offset(log(PRED1)) + PRED2 * PRED3 + PRED4 + 
(1|RANEF), family = "poisson", data = data, 
control=glmerControl(optimizer="bobyqa", optCtrl=list(maxfun=100000)))

summary(MOD1)

Generalized linear mixed model fit by maximum likelihood (Laplace 
Approximation) ['glmerMod'] Family: poisson ( log ) Formula: RESP ~ 
offset(log(PRED1)) + PRED2 * PRED3 + PRED4 + (1 | RANEF) Data: data 
Control: glmerControl(optimizer = "bobyqa", optCtrl = list(maxfun = 
1e+05)) AIC BIC logLik deviance df.resid 2451.3 2470.9 -1219.6 2439.3 
190 Scaled residuals: Min 1Q Median 3Q Max -10.8175 -1.2477 -0.1913 
1.1076 11.0389 Random effects: Groups Name Variance Std.Dev. RANEF 
(Intercept) 0.1519 0.3898 Number of obs: 196, groups: Order, 56 Fixed 
effects: Estimate Std. Error z value Pr(>|z|) (Intercept) -3.29896 
0.06250 -52.784 < 2e-16 *** PRED2 0.09089 0.01850 4.912 9.00e-07 *** 
PRED3 0.57262 0.02375 24.114 < 2e-16 *** PRED4 0.54726 0.01808 30.275 < 
2e-16 *** PRED2:PRED3 0.12190 0.01821 6.693 2.19e-11 *** --- Signif. 
codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 Correlation of 
Fixed Effects: (Intr) PRED2 PRED3 PRED4 PRED2 -0.142 PRED3 -0.056 0.102 
PRED4 -0.042 0.135 -0.155 PRED2:PRED3 0.001 0.017 -0.678 0.273


MODEL2 -  MCMCglmm using the lme4-package:

Now I build the same model using a MCMCglmm using weak informative 
priors. The offset term was implemented following the suggestion by 
Jarrod Hadfield in this post 
(https://stat.ethz.ch/pipermail/r-sig-mixed-models/2017q3/026016.html). 
See the model formula and summary below:


prior <- list(B=list(V=diag(6)*10, mu=c(0,1,0,0,0,0)),
                          G =list(G1=list(V=1,nu=0.002)),
                          R=list(V=1,nu=0.002))
prior$B$V[2,2] <- 1e-9 # set fixed effect prior for the offset variable to 1
prior


PGLMM.MOD1 <- MCMCglmm(RESP ~ log(PRED1) + PRED2 * PRED3 + PRED4,
                                         random = ~ RANEF,
                                         family="poisson",
                                         prior = prior,
                                         scale = T,
                                         pr=TRUE,
                                         data = data,
                                         nitt = 500000, burnin = 20000, 
thin = 200)


summary(PGLMM.MOD1)

Iterations = 20001:499801 Thinning interval = 200 Sample size = 2400 
DIC: 1164.591 G-structure: ~Order post.mean l-95% CI u-95% CI eff.samp 
RANEF 0.02264 0.0002387 0.0762 2400 R-structure: ~units post.mean l-95% 
CI u-95% CI eff.samp units 0.2788 0.1946 0.3705 2400 Location effects: 
RESP ~ log(PRED1) + PRED2 * PRED3 + PRED4 post.mean l-95% CI u-95% CI 
eff.samp pMCMC (Intercept) -3.40280 -3.50958 -3.28826 2400 <4e-04 *** 
log(PRED1) 1.00000 0.99994 1.00006 2257 <4e-04 *** PRED2 0.08369 
-0.01995 0.18120 2400 0.113 PRED3 0.63058 0.52164 0.73889 2400 <4e-04 
*** PRED4 0.60488 0.50565 0.69963 2400 <4e-04 *** PRED2:PRED3 0.06882 
-0.02363 0.17414 2258 0.167 --- Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 
‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1


As you can see from the different summaries, the estimates are fairly 
similar with some divergence especially regarding PRED2 and the 
interaction term. Additionally the provided p-values differ 
dramatically, with the GLMM (lme4) being highly significant for PRED2 
and the interaction wheras non-significance for the MCMCglmm.

My questions now are:

1. How can the divergence in results be explained and why doe both 
methods provide different results even though the model structure is the 
same? (Note: diagnostic plots for both models look fine)

2. I am aware that there is strong debate about the validity of p-values 
within GLMMs. Is there similar concern regarding MCMCglmm?

3. What would be a sensible interpretation of both model outputs 
especially with respect to PRED2 and the interaction term.


Any insight and help on that matter would be highly appreciated.

Best regards,

Bernd


-- 
Bernd Lenzner PhD

Researcher
University of Vienna
Department of Botany and Biodiversity Research
Division of Conservation Biology, Vegetation and Landscape Ecology
Rennweg 14
A-1030 Vienna



	[[alternative HTML version deleted]]



More information about the R-sig-mixed-models mailing list