experiments.bench
performs a
number of experiments sensitivity test.
It makes use of five different big gene datasets subsampling them to
generate different datasets.num of the network with different number
of experiments.experiments.bench(methods = "all.fast", datasources.names = "all",
experiments = c(20, 50, 150), eval = "AUPR",
no.topedges = 20, datasets.num = 3, local.noise = 20,
global.noise = 0, noiseType = "normal", sym = TRUE,
seed = NULL, verbose= TRUE)
evaluate
.datasource.subsample
.datasource.subsample
.datasource.subsample
.evaluate
.NULL
) - see set.seed
.experiments.bench
returns a list with three elements:
methods
at each one
of the specifieddatasources.names
with different noise
intensities.datasets.num
between the best algorithm
and the others.methods
accepts "all.fast" and "all"
(case insensitive) as a parameters:
no.topedges
% of the possible links
inferred by each algorithm at each dataset.Two different types of noises are added independently:
local.noise
specifies the percentage for each
variable ($\pm 20 %$).global.noise
specifies the percentage of the mean standard
deviation of all the variables ($\pm 20 %$).noiseType
, it is possible
to choose between "normal" (rnorm
) and "lognormal"
(rlnorm
). The argument noiseType
can be a
single character, this specifies the same distribution for both "Local"
and "Global" noise, it also can be a vector of characters with two
elements, the former specifies the distribution of "Local" noise and
the later the distribution of "Global" noise.netbenchmark
, noise.bench
results <- experiments.bench(datasources.names="toy",
datasets.num=2,methods="all.fast",experiments=c(20,40))
Run the code above in your browser using DataLab