[R-meta] Dear Wolfgang

Michael Dewey ||@t@ @end|ng |rom dewey@myzen@co@uk
Tue Mar 31 13:28:10 CEST 2020


Indeed if the single time point studies have used different times Ju 
will probably want to do a meta-regression with time as a moderator so 
it would not matter too much if Ju chose any single value from the 
multiple time point studies. That would avoid the complexity of 
estimating the V matrix.

Michael

On 30/03/2020 19:42, Dr. Gerta Rücker wrote:
> Dear Ju,
> 
> Another (maybe simplistic) solution could be to use only one (e.g., 
> always the first) time point for those studies that report repeated 
> measurements. This can be justified because you wrote that the "large 
> majority of studies measure this over short term experiment and thus on 
> a single time point" - so, as I understand, you it is only a minority of 
> studies that causes the multiplicity issue.
> 
> A similar simplification is often done when including a small number of 
> cross-over trials into a meta-analysis of RCTs with parallel-group design.
> 
> Best,
> 
> Gerta
> 
> Am 30.03.2020 um 20:36 schrieb Nicky Welton:
>> If you are interested in modelling the time-course relationship then 
>> there is a new package in R to do this within a network meta-analysis 
>> framework (although it can also be used if there are only 2 
>> interventions):
>> https://cran.r-project.org/web/packages/MBNMAtime/index.html
>>
>> Best wishes,
>>
>> Nicky
>>
>> -----Original Message-----
>> From: R-sig-meta-analysis <r-sig-meta-analysis-bounces using r-project.org> 
>> On Behalf Of Viechtbauer, Wolfgang (SP)
>> Sent: 30 March 2020 19:00
>> To: Ju Lee <juhyung2 using stanford.edu>; r-sig-meta-analysis using r-project.org
>> Subject: Re: [R-meta] Dear Wolfgang
>>
>> Thanks for the clarification.
>>
>> Computing a time-averaged d (or g) value is tricky because the values 
>> are not independent. So, if you meta-analyze them, the standard error 
>> of the pooled estimate would not be correct unless you take the 
>> dependency into consideration. And to answer one of your questions, 
>> squaring the standard error from the model would give you the sampling 
>> variance, but again, that value would not be correct.
>>
>> Basically what you have is the 'multiple-endpoint' case described here:
>>
>> http://www.metafor-project.org/doku.php/analyses:gleser2009#multiple-endpoint_studies 
>>
>>
>> You would need an estimate of the correlation between the repeated 
>> measurements (and if you have more than two time points, then an 
>> entire correlation matrix) to construct the V matrix for each study 
>> before meta-analyzing the values. Then you could use:
>>
>> res <- rma.mv(yi, V, data=dat)
>>
>> to pool the estimates into a time-averaged estimate, coef(res), and 
>> vcov(res) would give you the sampling variance. But the difficult part 
>> is constructing V.
>>
>> Maybe you can make some reasonable assumptions about the size of 
>> correlation (which probably should be lower the further time points 
>> are separated, although if there are seasonal effects, then 
>> measurements taken during similar seasons - even if they are further 
>> apart - may tend to be more correlated again). Based on equations 
>> (19.26) and (19.27) from the Gleser and Olkin  (2009) chapter, you can 
>> then construct the V matrix.
>>
>> Best,
>> Wolfgang
>>
>> -----Original Message-----
>> From: Ju Lee [mailto:juhyung2 using stanford.edu]
>> Sent: Monday, 30 March, 2020 18:16
>> To: Michael Dewey; Viechtbauer, Wolfgang (SP); 
>> r-sig-meta-analysis using r-project.org
>> Subject: Re: [R-meta] Dear Wolfgang
>>
>> Dear Wolfgang, Michael
>> These questions are important, and thank you for pointing them out.
>> In answers to your questions:
>> 1. Studies are measuring predation or herbivory rate on experimental 
>> prey in control and treatment conditions in the field. Large majority 
>> of studies measure this over short term experiment and thus on a 
>> single time point (ex. After 24 hr of field exposure).
>>
>> 2. However, some studies will monitor these responses over long-term 
>> period over multiple seasons to understand the seasonal dynamics. The 
>> issue here is that responses show seasonal fluctuation, which is what 
>> they were looking for. But the reviewer of our study has warned 
>> against using single representative timepoint within these multiple 
>> measures (ex. Peak season) but rather time-average these multiple 
>> measurements to reduce time-related bias.
>>
>> 3. So back to the point: In all of studies, in multiple time point, 
>> same treatment vs. control group are being compared using the same 
>> sampling method. We have mean and SD for all of these data point, 
>> separately for control and treatment groups for each time point of 
>> multiple measurements.
>>
>> 4. I am using Hedges' d as effect sizes.
>> Thank you and I hope this clarifies the question better!
>> Best,
>> JU
>>
>> ________________________________________
>> From: Michael Dewey <lists using dewey.myzen.co.uk>
>> Sent: Monday, March 30, 2020 5:32 AM
>> To: Viechtbauer, Wolfgang (SP) 
>> <wolfgang.viechtbauer using maastrichtuniversity.nl>; Ju Lee 
>> <juhyung2 using stanford.edu>; r-sig-meta-analysis using r-project.org 
>> <r-sig-meta-analysis using r-project.org>
>> Subject: Re: [R-meta] Dear Wolfgang
>> And in addition to Wolfgang's comments it would be helpful to know 
>> what scientific question underlies the decision to measure at multiple 
>> time points. Presumably the authors of primary studies did not do it 
>> for fun.
>>
>> Michael
>>
>> On 30/03/2020 11:37, Viechtbauer, Wolfgang (SP) wrote:
>>> Dear Ju,
>>>
>>> Before I can try to address your actual questions, please say a bit 
>>> more about the studies that measure responses at a single time point. 
>>> Are groups (e.g., treatment versus control) being compared within 
>>> these studies? Are the 'responses' continuous (such that means and 
>>> SDs are being reported) or dichotomous (such that counts or 
>>> proportions are being reported) or something else? And related to 
>>> this, what effect size measure are you using for quantifying the 
>>> group difference within studies? Standardized mean differences (which 
>>> would make sense when means/SDs are being reported), risk differences 
>>> or (log) risk/odds ratios (based on counts/proportions), or something 
>>> else?
>>>
>>> And the studies that measure responses at multiple time points: Are 
>>> they just doing the same thing that the 'single time point studies' 
>>> are doing, but at multiple time points? For example, instead of 
>>> reporting the means and SDs of the treatment and control group once, 
>>> there are several follow-ups, such that such the means and SDs of the 
>>> two groups are reported at each follow-up time point?
>>>
>>> Best,
>>> Wolfgang
>>>
>>> -----Original Message-----
>>> From: R-sig-meta-analysis
>>> [mailto:r-sig-meta-analysis-bounces using r-project.org] On Behalf Of Ju Lee
>>> Sent: Sunday, 29 March, 2020 20:16
>>> To: r-sig-meta-analysis using r-project.org
>>> Subject: [R-meta] Dear Wolfgang
>>>
>>> Dear Wolfgang,
>>>
>>> I sincerely hope you are well and healthy.
>>> I wanted to reach out regarding ways to incorporate studies with 
>>> repeated-measures to overall mixed effect models.
>>>
>>> My data is almost entirely composed of studies measuring responses at 
>>> a single time point, but there are few studies that have been 
>>> measuring responses multiple times throughout study seasons. I was 
>>> advised that time-averaging these multiple responses makes more sense 
>>> for these studies.
>>>
>>> My understanding was that you could 1) do a fixed effect 
>>> meta-analysis of these studies to generate a single mean effect sizes 
>>> and sampling variance from these repeated measurements and then 2) 
>>> incorporate the single effect size and variance into overall 
>>> mixed-effect model. Is this a correct approach?
>>>
>>> If so, how would I calculate sampling variance from the fixed model 
>>> in the step 1? Is it based on SE outputs of the fixed effect model?
>>>
>>> Thank you very much, and I look forward to hearing from you!
>>> Best,
>>> JU
>> _______________________________________________
>> R-sig-meta-analysis mailing list
>> R-sig-meta-analysis using r-project.org
>> https://stat.ethz.ch/mailman/listinfo/r-sig-meta-analysis
>>
>> _______________________________________________
>> R-sig-meta-analysis mailing list
>> R-sig-meta-analysis using r-project.org
>> https://stat.ethz.ch/mailman/listinfo/r-sig-meta-analysis
> 
> _______________________________________________
> R-sig-meta-analysis mailing list
> R-sig-meta-analysis using r-project.org
> https://stat.ethz.ch/mailman/listinfo/r-sig-meta-analysis
> 

-- 
Michael
http://www.dewey.myzen.co.uk/home.html



More information about the R-sig-meta-analysis mailing list