
Last chance! 50% off unlimited learning
Sale ends in
Used in the development of emulations of a simulation using a latin-hypercube summary file
partition_dataset(dataset, parameters, measures, percent_train = 75,
percent_test = 15, percent_validation = 10, seed = NULL,
normalise = FALSE, sample_mins = NULL, sample_maxes = NULL,
timepoint = NULL)
LHC summary file to partition
Simulation parameters the emulation will be fed as input
Simulation responses of interest
Percent of the dataset to use as training
Percent of the dataset to use as testing
Percent of the dataset to use as validation
For specifying a particular seed when randomly splitting the set
Whether the data needs to be normalised (to be between 0 and 1). For emulation creation to be successful, all data must be normalised prior to use in training and testing
The minimum value used for each parameter in generating the latin-hypercube sample
The maximum value used for each parameter in generating the latin-hypercube sample
Simulation timepoint for which this summary file was created
Partitioned dataset containing training, testing, and validation sets, in addition to the sample mins and maxes such that any predictions that are generated using this normalised data can be rescaled correctly
# NOT RUN {
data("sim_data_for_emulation")
parameters<-c("stableBindProbability","chemokineExpressionThreshold",
"initialChemokineExpressionValue","maxChemokineExpressionValue",
"maxProbabilityOfAdhesion","adhesionFactorExpressionSlope")
measures<-c("Velocity","Displacement","PatchArea")
sample_maxes <- cbind(100,0.9,0.5,0.08,1,5)
sample_mins <-cbind(0,0.1,0.1,0.015,0.1,0.25)
partitionedData <- partition_dataset(sim_data_for_emulation, parameters,
measures, percent_train=75, percent_test=15, percent_validation=10, normalise=TRUE,
sample_mins = sample_mins, sample_maxes = sample_maxes)
# }
Run the code above in your browser using DataLab