[R-sig-ME] Modeling truncated counts with glmer
Thierry Onkelinx
thierry.onkelinx at inbo.be
Mon Jan 23 09:42:01 CET 2017
Dear João,
A binomial distribution seems more relevant to me.
glmer(cbind(correctPair, incorrectPair) ~ I((abruf - 1)^2) * treatment +
(1|subjectNumber), data=data, family = binomial)
Best regards,
ir. Thierry Onkelinx
Instituut voor natuur- en bosonderzoek / Research Institute for Nature and
Forest
team Biometrie & Kwaliteitszorg / team Biometrics & Quality Assurance
Kliniekstraat 25
1070 Anderlecht
Belgium
To call in the statistician after the experiment is done may be no more
than asking him to perform a post-mortem examination: he may be able to say
what the experiment died of. ~ Sir Ronald Aylmer Fisher
The plural of anecdote is not data. ~ Roger Brinner
The combination of some data and an aching desire for an answer does not
ensure that a reasonable answer can be extracted from a given body of data.
~ John Tukey
2017-01-23 8:46 GMT+01:00 João C P Santiago <joao.santiago op uni-tuebingen.de>
:
> Hi,
>
> In my experiment 20 participants did a word-pairs learning task in two
> conditions (repeated measures):
> 40 pairs of nouns are presented on a monitor, each for 4s and with an
> interval of 1s. The words of each pair were moderately semantically related
> (e.g., brain, consciousness and solution, problem). Two different word
> lists were used for the subject’s two experimental conditions, with the
> order of word lists balanced across subjects and conditions. The subject
> had unlimited time to recall the appropriate response word, and did three
> trials in succession for each list:
>
> Condition 1, List A > T1, T2, T3
> Condition 2, List B > T1, T2, T3
>
> No feedback was given as to whether the remembered word was correct or not.
>
> I've seen some people go at this with anova, others subtract the total
> number of correct pairs in one condition from the other per subject and run
> a t-test. Since this is count data, a generalized linear model should be
> more appropriate, right?
>
> head(data)
> subjectNumber expDay bmi treatment tones hour abruf
> correctPair incorrectPair
> <dbl> <chr> <dbl> <fctr> <dbl> <time> <dbl>
> <dbl> <dbl>
> 1 1 N2 22.53086 Control 0 27900 secs 1
> 26 14
> 2 1 N2 22.53086 Control 0 27900 secs 2
> 40 0
> 3 1 N2 22.53086 Control 0 27900 secs 3
> 40 0
> 4 2 N1 22.53086 Control 0 27900 secs 1
> 22 18
> 5 2 N1 22.53086 Control 0 27900 secs 2
> 33 7
> 6 2 N1 22.53086 Control 0 27900 secs 3
> 36 4
>
>
>
> I fitted a model with glmer.nb(correctPair ~ I((abruf - 1)^2) * treatment
> + (1|subjectNumber), data=data). The residuals don't look so good to me
> http://imgur.com/a/AJXGq and the model is fitting values above 40, which
> will never happen in real life (not sure if this is important).
>
> I'm interested in knowing if there is any difference between conditions
> (are the values at timepoint (abruf) 1 different? do people remember less
> in one one condition than in the other (different number of pairs at
> timepoint 3?)
>
>
> If the direction I'm taking is completely wrong please let me know.
>
> Best,
> Santiago
>
>
>
> --
> João C. P. Santiago
> Institute for Medical Psychology & Behavioral Neurobiology
> Center of Integrative Neuroscience
> University of Tuebingen
> Otfried-Mueller-Str. 25
> 72076 Tuebingen, Germany
>
> Phone: +49 7071 29 88981
> Fax: +49 7071 29 25016
>
> _______________________________________________
> R-sig-mixed-models op r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-sig-mixed-models
[[alternative HTML version deleted]]
More information about the R-sig-mixed-models
mailing list