[R-meta] Meta-analysis of within-subject experimental designs with multiple treatment factors?
Schlüter, Elmar
E|m@r@Sch|ueter @end|ng |rom @ow|@un|-g|e@@en@de
Fri Sep 15 09:12:45 CEST 2023
Dear James,
I would just like to say "thank you" for your quick and helpful reply - I will be happy to try and adapt my models accordingly.
Best,
Elmar
________________________________
Von: James Pustejovsky <jepusto using gmail.com>
Gesendet: Donnerstag, 14. September 2023 16:22:15
An: R Special Interest Group for Meta-Analysis
Cc: Schlüter, Elmar
Betreff: Re: [R-meta] Meta-analysis of within-subject experimental designs with multiple treatment factors?
Hi Elmar,
The example code you provided checks out as far as I can see. Your approach seems reasonable to me, although I also see two ways that it could be refined:
* You've assumed a constant correlation of 0.6 between the sampling errors of the effect size estimates from a given study. This assumption could potentially be refined depending on how much information you have about the study results. If you had access or could gain access to the raw data, you could calculate better estimates of the variance-covariance matrix of these effect estimates.
* You've used what I call the "correlated-and-hierarchical effects" working model, which includes random effects for each individual effect and random effects for each study. This working model assumes that 1) the overall variation in each type of effect is constant across all four types (i.e., between-study heterogeneity of slope_A is the same as the between-study heterogeneity of slope_B, likewise for slope_C1 and slope_C2) and that 2) there is a constant correlation between the study-specific true effects of each pair of effect types--that is, cor(slope_A, slope_B) is the same as cor(slope_A, slope_C1) and cor(slope_A, slope_C2) and cor(slope_B, slope_C1), etc. If you have a fair number studies, you could consider using a more general, multivariate working model, which would loosen both of the above assumptions. The syntax is just as above, but with random effects specified as
random = ~ coeff_name | study, struct = "UN"
James
On Thu, Sep 14, 2023 at 4:55 AM Schlüter, Elmar via R-sig-meta-analysis <r-sig-meta-analysis using r-project.org<mailto:r-sig-meta-analysis using r-project.org>> wrote:
Dear Listmembers,
here is a repost of my earlier mail that I unfortunately posted in html-format... I hope this time the data example is readable:
---
I am planning to conduct a meta-analysis based on data from a series of factorial survey studies. The factorial survey studies I'd like to analyze correspond to within-subject experimental designs with multiple treatment factors. However, there seem to be very few, if any, meta-analyses using data from factorial survey studies or related designs. I thus lack examples corresponding to the design I need to ask some questions about below. So here's a brief description of my data in general terms:
A factorial survey, also known as a vignette study or factorial vignette design, is a research method used to study how individuals make judgments or decisions in complex situations by presenting them with a series of hypothetical scenarios (vignettes) that vary systematically across multiple factors or attributes. For example, respondents might be asked how likely they would vote for a candidate of a political party who is
- male vs. female, (factor 1)
- has a leftwing vs. rightwing orientation (factor 2)
- and behaves very empathic/somewhat empathic/not empathic (factor 3).
This results in a 2 x 2 x 3 factorial design with 12 different vignettes.
In the original data, the factor levels were measured using dummy variables. The results are unstandardised regression slopes plus standard errors. Accordingly, I'd like to meta-analyse 4 unstandardised regression slopes:
- the slope of the dummy variable for factor A (the average effect of male vs. female), fA_slope
- the slope of the dummy variable for factor B (the average effect of rightwing vs. leftwing orientation), fB_slope
- the two slopes of the two dummy variables for factor C (somewhat empathic vs very empathic and not empathic vs very empathic), fC_slope1 & fC_slope2
Using one dependent variable, all studies were conducted as full or fractional factorial within-subjects designs. Therefore, I believe it would be best to account for any potential non-independence in the data, given that each respondent rated multiple (up to 12) vignettes. What I find a bit challenging is that there is not just one average effect to examine, but four in total. Furthermore, these four average slopes result from three orthogonal factors. I (think I) have been able to combine clubSandwich and metafor to examine these data - but I would be very happy to hear if you think my approach is adequate, or if there are alternative ways to analyse data from such a research design?
Here's an illustration/sample of the data restricted to 5 studies only, the real dataset will comprise of approx. 20 studies. Notice that I arranged all coefficient values in one column:
Data (df) for illustration:
study coeff_name coeff_value v coeff_index
A fA_slope1 -0.07 0.000113 1
A fB_slope1 0.17 0.000344 2
A fC_slope1 -0.08 0.000160 3
A fC_slope2 -0.12 0.000171 4
B fA_slope1 -0.08 0.000179 5
B fB_slope1 0.18 0.000504 6
B fC_slope1 -0.06 0.000309 7
B fC_slope2 -0.14 0.000435 8
C fA_slope1 -0.06 0.000097 9
C fB_slope1 0.16 0.000319 10
C fC_slope1 -0.07 0.000150 11
C fC_slope2 -0.11 0.000152 12
D fA_slope1 -0.03 0.000107 13
D fB_slope1 0.15 0.000452 14
D fC_slope1 -0.05 0.000198 15
D fC_slope2 -0.10 0.000196 16
E fA_slope1 -0.04 0.000123 17
E fB_slope1 0.15 0.000366 18
E fC_slope1 -0.11 0.000193 19
E fC_slope2 -0.18 0.000205 20
Here's the code I used:
library(metafor)
library(clubSandwich)
# Create sampling variance covariance matrix
V_mat <- impute_covariance_matrix(Testdata$v, cluster = df$study, r = 0.6)
# model in metafor
mod <- rma.mv<http://rma.mv>(coeff_value ~ 0 + coeff_name, V = V_mat, #coeff_name contains all coefficients as factor levels
random = ~ 1 | study / coeff_index, #coefficients are nested in studies
data = df)
# clustered SEs and CIs
conf_int(mod, vcov = "CR2")
And here are the results:
Coef. Estimate SE d.f. Lower_95%CI Upper_95%CI
coeff_namecfA_slope1 -0.0553 0.00924 3.99 -0.081 -0.0296
coeff_namecfB_slope1 0.1623 0.0057 3.97 0.146 0.1782
coeff_namecfC_slope1 -0.0739 0.01044 3.97 -0.103 -0.0448
coeff_namecfC_slope2 -0.1289 0.01465 3.92 -0.17 -0.0879
The code seems to work. But as I outlined at the beginning, it is not clear to if one might arrange the data in "long format" with all coefficients of interest in one column as I did. I thus highly appreciate your advice and any comments
Best,
Elmar
[[alternative HTML version deleted]]
_______________________________________________
R-sig-meta-analysis mailing list @ R-sig-meta-analysis using r-project.org<mailto:R-sig-meta-analysis using r-project.org>
To manage your subscription to this mailing list, go to:
https://stat.ethz.ch/mailman/listinfo/r-sig-meta-analysis
[[alternative HTML version deleted]]
More information about the R-sig-meta-analysis
mailing list