step_tfidf
creates a specification of a recipe step that
will convert a tokenlist into multiple variables containing
the term frequency-inverse document frequency of tokens.
step_tfidf(
recipe,
...,
role = "predictor",
trained = FALSE,
columns = NULL,
vocabulary = NULL,
res = NULL,
smooth_idf = TRUE,
norm = "l1",
sublinear_tf = FALSE,
prefix = "tfidf",
skip = FALSE,
id = rand_id("tfidf")
)# S3 method for step_tfidf
tidy(x, ...)
A recipe object. The step will be added to the sequence of operations for this recipe.
One or more selector functions to choose variables.
For step_tfidf
, this indicates the variables to be encoded
into a tokenlist. See recipes::selections()
for more
details. For the tidy
method, these are not currently used.
For model terms created by this step, what analysis role should they be assigned?. By default, the function assumes that the new columns created by the original variables will be used as predictors in a model.
A logical to indicate if the recipe has been baked.
A list of tibble results that define the
encoding. This is NULL
until the step is trained by
recipes::prep.recipe()
.
A character vector of strings to be considered.
The words that will be used to calculate the term
frequency will be stored here once this preprocessing step has
be trained by prep.recipe()
.
TRUE smooth IDF weights by adding one to document frequencies, as if an extra document was seen containing every term in the collection exactly once. This prevents division by zero.
A character, defines the type of normalization to apply to term vectors. "l1" by default, i.e., scale by the number of words in the document. Must be one of c("l1", "l2", "none").
A logical, apply sublinear term-frequency scaling, i.e., replace the term frequency with 1 + log(TF). Defaults to FALSE.
A character string that will be the prefix to the resulting new variables. See notes below.
A logical. Should the step be skipped when the
recipe is baked by recipes::bake.recipe()
? While all
operations are baked when recipes::prep.recipe()
is run, some
operations may not be able to be conducted on new data (e.g.
processing the outcome variable(s)). Care should be taken when
using skip = TRUE
as it may affect the computations for
subsequent operations.
A character string that is unique to this step to identify it.
A step_tfidf
object.
An updated version of recipe
with the new step added
to the sequence of existing steps (if any).
It is strongly advised to use step_tokenfilter before using step_tfidf to limit the number of variables created; otherwise you may run into memory issues. A good strategy is to start with a low token count and increase depending on how much RAM you want to use.
Term frequency-inverse document frequency is the product of two statistics: the term frequency (TF) and the inverse document frequency (IDF).
Term frequency measures how many times each token appears in each observation.
Inverse document frequency is a measure of how informative a word is, e.g., how common or rare the word is across all the observations. If a word appears in all the observations it might not give that much insight, but if it only appears in some it might help differentiate between observations.
The IDF is defined as follows: idf = log(1 + (# documents in the corpus) / (# documents where the term appears))
The new components will have names that begin with prefix
, then
the name of the variable, followed by the tokens all separated by
-
. The new variables will be created alphabetically according to
token.
step_tokenize()
to turn character into tokenlist.
Other tokenlist to numeric steps:
step_texthash()
,
step_tf()
,
step_word_embeddings()
# NOT RUN {
library(recipes)
library(modeldata)
data(okc_text)
okc_rec <- recipe(~ ., data = okc_text) %>%
step_tokenize(essay0) %>%
step_tfidf(essay0)
okc_obj <- okc_rec %>%
prep()
bake(okc_obj, okc_text)
tidy(okc_rec, number = 2)
tidy(okc_obj, number = 2)
# }
Run the code above in your browser using DataLab