[R-meta] Assessing selection bias / multivariate meta-analysis
James Pustejovsky
jepu@to @end|ng |rom gm@||@com
Thu Nov 21 15:10:00 CET 2024
Was going to chime in about the metaselection package (
https://github.com/jepusto/metaselection)---it's still under development
but the core functionality and documentation is in place. The package
implements the bootstrapped selection model as demonstrated in my blog post
(https://jepusto.com/posts/cluster-bootstrap-selection-model/), but with a
much easier interface and faster calculation; it also implements selection
models with cluster-robust standard errors, though these seem to be not as
accurate as bootstrapping. Folks are welcome to give the package a try and
to reach out with questions or potential bugs if you run into anything. We
are working on a paper describing the methods implemented in the package
and reporting pretty extensive simulations about their performance.
My student Man Chen (now on the faculty at UT Austin) has studied a whole
bunch of the available methods for selective reporting bias correction,
looking specifically at how they perform in meta-analyses with dependent
effect sizes, and proposing adaptations of some of the methods to better
acknowledge dependency. Our working paper (on this is here:
https://osf.io/preprints/metaarxiv/jq52s
Pia asked about a few other possible techniques:
- The Egger's test / PET-PEESE approach with cluster-robust variance
estimation is reasonable but, as Wolfgang noted, it is not specifically
diagnostic about missing studies vs. missing effects. If the effect sizes
nested within a given study tend to have similar standard errors, then it
will mostly be picking up on association between study sample size and
study-level average effect size. And of course, it also has the limitation
that this small-study association can be caused by things other than
selective reporting.
- Mathur & Vanderweele's sensitivity analysis is quite useful, though it
does not provide an estimate of the severity of selective reporting
(instead, it provides information about the degree of potential bias
assuming a specific level of selection).
- For 3PSM, the cluster-bootstrap technique implemented in the
metaselection package is a way to deal with dependent effects, so it is no
longer necessary to use ad hoc approaches like ignoring dependence,
aggregating to the study level, or selecting a single effect per study.
James
On Thu, Nov 21, 2024 at 6:37 AM Viechtbauer, Wolfgang (NP) via
R-sig-meta-analysis <r-sig-meta-analysis using r-project.org> wrote:
> And I just stumbled across this:
>
> https://github.com/jepusto/metaselection
>
> James, don't hide all your good work from us!
>
> Best,
> Wolfgang
>
> > -----Original Message-----
> > From: R-sig-meta-analysis <r-sig-meta-analysis-bounces using r-project.org>
> On Behalf
> > Of Viechtbauer, Wolfgang (NP) via R-sig-meta-analysis
> > Sent: Thursday, November 21, 2024 13:21
> > To: R Special Interest Group for Meta-Analysis <r-sig-meta-analysis using r-
> > project.org>
> > Cc: Viechtbauer, Wolfgang (NP) <
> wolfgang.viechtbauer using maastrichtuniversity.nl>
> > Subject: Re: [R-meta] Assessing selection bias / multivariate
> meta-analysis
> >
> > Dear Pia,
> >
> > Generally, I don't think there really is any method that is going to be
> a great
> > choice here. The 'Egger sandwich' (i.e., an Egger type regression model
> using
> > cluster-robust inference methods) is a decent option, since it logically
> > generalizes the standard Egger regression method to this context, but it
> is
> > unclear what kind of bias/selection effect this may pick up (missing
> studies,
> > missing estimates within studies, a combination thereof).
> >
> > Yes, for the 3PSM, you would have to either ignore the dependencies or
> select
> > one estimate per study (and maybe repeat the latter a large number of
> times for
> > different subsets).
> >
> > I assume you are familiar with these papers. If not, they are directly
> relevant:
> >
> > Rodgers, M. A., & Pustejovsky, J. E. (2021). Evaluating meta-analytic
> methods to
> > detect selective reporting in the presence of dependent effect sizes.
> > Psychological Methods, 26(2), 141-160.
> https://doi.org/10.1037/met0000300
> >
> > Fernández-Castilla, B., Declercq, L., Jamshidi, L., Beretvas, S. N.,
> Onghena,
> > P., & Van den Noortgate, W. (2021). Detecting selection bias in
> meta-analyses
> > with multiple outcomes: A simulation study. The Journal of Experimental
> > Education, 89(1), 125-144. https://doi.org/10.1080/00220973.2019.1582470
> >
> > Nakagawa, S., Lagisz, M., Jennions, M. D., Koricheva, J., Noble, D. W.
> A.,
> > Parker, T. H., Sánchez-Tójar, A., Yang, Y., & O'Dea, R. E. (2022).
> Methods for
> > testing publication bias in ecological and evolutionary meta-analyses.
> Methods
> > in Ecology and Evolution, 13(1), 4-21.
> https://doi.org/10.1111/2041-210X.13724
> >
> > I think James is working on some methods related to this topic:
> >
> > https://jepusto.com/posts/cluster-bootstrap-selection-model/
> >
> > Best,
> > Wolfgang
> >
> > > -----Original Message-----
> > > From: R-sig-meta-analysis <r-sig-meta-analysis-bounces using r-project.org>
> On
> > Behalf
> > > Of Pia-Magdalena Schmidt via R-sig-meta-analysis
> > > Sent: Wednesday, November 20, 2024 21:58
> > > To: r-sig-meta-analysis using r-project.org
> > > Cc: Pia-Magdalena Schmidt <pia-magdalena.schmidt using uni-bonn.de>
> > > Subject: [R-meta] Assessing selection bias / multivariate meta-analysis
> > >
> > > Dear all,
> > > Although this topic has been discussed several times and I read the
> archives
> > > and referenced papers, I’m still not sure how to assess and possibly
> correct
> > > for selection bias in multivariate meta-analyses.
> > >
> > > I used the metafor package and ran meta-analyses with SMCC as effect
> size
> > > (all studies used within-designs) and fitted rma.mv models as several
> > > studies report more than one effect size. Furthermore, I used
> cluster-robust
> > > methods to examine the robustness of the models.
> > > For a subset of my data, I used meta-regressions with one continuous
> > > moderator.
> > > All effect sizes are from published journal articles. The range of
> included
> > > studies is between 30 and 6 with a number of effect sizes between 45
> and 10.
> > >
> > > Since I want to take the dependencies into account, I would not use
> funnel
> > > plots or trim and fill. I wonder if using Egger's regression test
> adjusted
> > > for rma.mv as well as PET-PEESE and perhaps the sensitivity analysis
> > > suggested by Mathur & Vanderweele (2020) as well as 3PSM would be a
> > > reasonable way to go? Although the latter would only use one effect
> size per
> > > study or an aggregated effect size, right?
> > >
> > > I would be very grateful for any recommendations!
> > > Best,
> > > Pia
> > >
> > > Below is an excerpt from my code:
> > > ES_all <- escalc(measure="SMCC", m1i= m1i, sd1i= sd1i, ni = ni, m2i=
> m2i,
> > > sd2i= sd2i, pi= pi, ri = ri, data= dat)
> > > V <- vcalc(vi=ES_all$vi, cluster=id_database, obs = effect_id, rho
> =0.605,
> > > data=dat)
> > > res <- rma.mv(yi=ES_all$yi, V, random = ~ 1 | id_database/effect_id,
> data =
> > > dat)
> > > res.robust <- robust(res, cluster = id_database, clubSandwich = TRUE)
> > >
> > > # subset
> > > res_LOR <- rma.mv(yi=ES_LOR$yi, V, random = ~ 1 |
> id_database/effect_id,
> > > mods = ~ dose, data = dat)
> _______________________________________________
> R-sig-meta-analysis mailing list @ R-sig-meta-analysis using r-project.org
> To manage your subscription to this mailing list, go to:
> https://stat.ethz.ch/mailman/listinfo/r-sig-meta-analysis
>
[[alternative HTML version deleted]]
More information about the R-sig-meta-analysis
mailing list