[BioC] Re: Use of RMA in increasingly-sized datasets
Swati Ranade
ssr3072 at yahoo.com
Mon Jun 6 15:27:21 CEST 2005
Hi Kawai,
I think you might be able to generate a dummy training
set from Published microarray data!!
Swati
-------------------------------------------------------
I am in just the same problem as yours now.
I think there are two key steps in the RMA that depend
on the set of
chips
in a run. One is quantile normalization step, and the
other is median
polish summarization step. The target value of each
quantile of probe
intensity is the geometrical mean calculated from the
same qunatiles
across
the entire chip set in the run. And the expression
values summarized
from
11-20 probe intensities are calculated from median
polish algorithm
using
the probe sets across the entire chip set.
Therefore, the suggestion of the usage of "a standard
training 50 chip
set"
is effective in practice, because the fluctuation of
quantile target
value
is quite a little after adding one chip data to 50
chip standard set,
and
the median values used in the summarization step are
robust enough for
the
51 chip data set.
But this method is very tedious when we process
several chip data one
by
one, and to create the standard set is impossible at
the beginning of a
project.
I am looking forward to hearing some good solution on
this problem,
too.
Bye.
Kawai
More information about the Bioconductor
mailing list