Downloads pages from Wikipedia and extracts some meta information
with functions from the package WikipediR
. Creates a
textmeta
object including the requested pages.
readWiki(
category,
subcategories = TRUE,
language = "en",
project = "wikipedia"
)
textmeta
object
character
articles of which category should be
downloaded, see pages_in_category
, argument categories
logical
(default: TRUE
) should
subcategories be downloaded as well
character
(default: "en"
),
see pages_in_category
character
(default: "wikipedia"
),
see pages_in_category
if (FALSE) corpus <- readWiki(category="Person_(Studentenbewegung)",
subcategories = FALSE, language = "de", project = "wikipedia")
Run the code above in your browser using DataLab