[R-meta] negative reliability

Michael Dewey ||@t@ @end|ng |rom dewey@myzen@co@uk
Sat Apr 29 11:47:45 CEST 2023


I think I would just include the one where they use the same method as 
other authors. That seems simpler and avoids inroducing unnecessary 
heterogeneity.

Michael

On 28/04/2023 23:25, Catia Oliveira wrote:
> Dear Michael and Wolfgang,
> 
> Thank you for your reply. I am also intrigued about why a negative 
> correlation between test and retest would be encountered. I will think 
> carefully and plan to do sensitivity analysis.
> Do you have any thoughts about the second question?
> 
> A second issue, but somewhat in line with the previous one, what do you 
> recommend one to do when multiple approaches are used to compute the 
> reliability of the task but only one converges with what was typically 
> done by other authors? I wouldn't be able to assess whether the 
> decisions made an impact on the reliability as it is only one study but 
> also don't want to bias the findings with my selection (though I have to 
> say the results are quite consistent across approaches). Or do you think 
> I should just include all the results as I am already using a multilevel 
> approach? I just wouldn't be able to test whether the manipulations 
> affect the results and may be increasing the heterogeneity of the results.
> 
> Thank you!
> 
> Catia
> 
> On Fri, 28 Apr 2023 at 11:44, Viechtbauer, Wolfgang (NP) 
> <wolfgang.viechtbauer using maastrichtuniversity.nl 
> <mailto:wolfgang.viechtbauer using maastrichtuniversity.nl>> wrote:
> 
>     I am also not familiar with the correction by Krus and Helmstadter
>     (1993) (seems to be this article in case anybody is interested:
>     https://doi.org/10.1177/0013164493053003005
>     <https://doi.org/10.1177/0013164493053003005>) so I cannot really
>     comment on this. I think in the end, you just have to make your own
>     decisions here (unless somebody comes with further wisdom) and
>     document your choices. As always, a sensitivity analysis is also an
>     option.
> 
>     P.S.: Your message also arrived in my Inbox (if you recall from our
>     previous communication, this was an issue in the past, but with the
>     adjusted settings to the mailing list, this now seems to be resolved).
> 
>     Best,
>     Wolfgang
> 
>      >-----Original Message-----
>      >From: Michael Dewey [mailto:lists using dewey.myzen.co.uk
>     <mailto:lists using dewey.myzen.co.uk>]
>      >Sent: Friday, 28 April, 2023 11:01
>      >To: R Special Interest Group for Meta-Analysis; Viechtbauer,
>     Wolfgang (NP); James
>      >Pustejovsky
>      >Cc: Catia Oliveira
>      >Subject: Re: [R-meta] negative reliability
>      >
>      >Dear Catia
>      >
>      >You can check whether it was transmitted by going to
>      >
>      >https://stat.ethz.ch/pipermail/r-sig-meta-analysis/2023-April/author.html <https://stat.ethz.ch/pipermail/r-sig-meta-analysis/2023-April/author.html>
>      >
>      >Where it appears.
>      >
>      >The fact that you got no response may be because we are all struggling
>      >with the idea of a test-retest or split-half reliability estimate
>     which
>      >was negative and what we would do with it. So people who scored
>     high the
>      >first time now score low? If it is split-half it suggests that the
>      >hypothesis that the test measures one thing is false.
>      >
>      >Michael
>      >
>      >On 28/04/2023 01:45, Catia Oliveira via R-sig-meta-analysis wrote:
>      >> Dear all,
>      >>
>      >> I apologise if I am spamming you but I think you didn't receive
>     my previous
>      >> email. At least I was not notified.
>      >>
>      >> I am running a meta-analysis on the reliability of a task
>     (computed as a
>      >> correlation between sessions or halves of the task depending on
>     whether it
>      >> is test-retest or split-half reliability) and I have come across
>     one result
>      >> that I am not sure how to handle. According to the authors, they
>     found
>      >> negative reliability and, because of that, they applied a correction
>      >> suggested by Krus and Helmstadter(1993). Thus, I am wondering if
>     I should
>      >> use the original correlation or the corrected one. When authors
>     applied the
>      >> Spearman-Brown correction I reverted them to the original score,
>     but with
>      >> this one I don't know if such an approach is OK. My intuition
>     would be to
>      >> use the uncorrected measure since that's the most common
>     approach in the
>      >> sample and there isn't sufficient information to allow us to
>     test the
>      >> impact of these corrections. But I would appreciate your input
>     on this.
>      >>
>      >> A second issue, but somewhat in line with the previous one, what
>     do you
>      >> recommend one to do when multiple approaches are used to compute the
>      >> reliability of the task but only one converges with what was
>     typically done
>      >> by other authors? I wouldn't be able to assess whether the
>     decisions made
>      >> an impact on the reliability as it is only one study but also
>     don't want to
>      >> bias the findings with my selection (though I have to say the
>     results are
>      >> quite consistent across approaches).
>      >>
>      >> Thank you.
>      >>
>      >> Best wishes,
>      >>
>      >> Catia
> 
> 
> <http://www.avg.com/email-signature?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=emailclient>	Virus-free.www.avg.com <http://www.avg.com/email-signature?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=emailclient>
> 
> <#DAB4FAD8-2DD7-40BB-A1B8-4E2AA1F9FDF2>

-- 
Michael
http://www.dewey.myzen.co.uk/home.html



More information about the R-sig-meta-analysis mailing list