Learn R Programming

volkeR-Package

High-level functions for tabulating, charting and reporting survey data.

Getting started

# Install the package (see below), then load it
library(volker)

# Load example data from the package
data <- volker::chatgpt

# Create your first table and plot, counting answers to an item battery
report_counts(data, starts_with("cg_adoption_social"))

# Create your first table and plot, reporting mean values of the item battery
report_metrics(data, starts_with("cg_adoption_social"))

See further examples in vignette("introduction", package="volker").

Don’t miss the template feature: Within RStudio, create a new Markdown document, select From template, choose and finally knit the volkeR Report! It’s a blueprint for your own tidy reports.

Concept

The volkeR package is made for creating quick and easy overviews about datasets. It handles standard cases with a handful of functions. Basically you select one of the following functions and throw your data in:

  • Categorical variables: report_counts()
  • Metric variables: report_metrics()

The report functions combine tables, plots and, optionally, effect size calculations. To request only one of those outputs, directly use the respective function:

  • Charts: plot_metrics() and plot_counts()
  • Tables: tab_metrics() and tab_counts()
  • Effects: effect_metrics() and effect_counts()

Which one is best? That depends on your objective:

  • Table or plot?
    A plot is quick to capture, data from a table is better for further calculations. Functions for tables start with tab, functions for plots with plot. If in doubt, create both at once with the report-functions.

  • Categorical or metric variables?
    Categories can be counted, for metric variables distribution parameters such as the mean and standard deviation are calculated. Functions for categorical variables contain counts in their name, those for metric metrics.

  • Individual, grouped or correlated?
    Groups can be compared (e.g., the average age by gender) or cross-tabulated (e.g. combinations of education level and gender) by providing a grouping column as third parameter of table, plot and report functions. To calculate correlations and show scatter plots, provide a metric column and set the metric-Paramter to TRUE. The effect-functions calculate effect sizes and statistical tests for group comparisons and correlations.

  • One variable or item batteries?.
    Item batteries are often used in surveys. Each item results in a single variable, but the variables are all measured with the same scale (e.g. 1 = not at all to 5 = fully applies). To summarise multiple items send a column selection to the functions by using tidyselect mechanisms such as starts_with().

  • Markdown or data frame?
    All table functions return data frames that can be processed further. The tables have their own print function, so the output of all functions can be used directly in Markdown documents to display neatly formatted tables and plots. The report-functions create tidy interactive tabsheets to switch between plots, tables, and indexes.

Examples

All functions take a data frame as their first argument, followed by a column selection, and optionally a grouping column. Reproduce the examples above:

  • One metric variable: report_metrics(data, sd_age)
  • One categorical variable: report_counts(data, sd_gender)
  • Grouped metric variable: report_metrics(data, sd_age, sd_gender)
  • Grouped categorical variable: report_counts(data, adopter, sd_gender)
  • Multiple metric variables: report_metrics(data, starts_with("cg_adoption"))
  • Multiple categorical variables: report_counts(data, starts_with("cg_adoption"))

The column selections determine which type of output is generated. In the second parameter (after the dataset), you can either provide a single column or a selection of multiple items. To compare groups, provide an additional categorical column in the third parameter. To calculate correlations, provide a metric column in the third parameter and set the metric parameter to TRUE.

Note: Some column combinations are not implemented yet.

Effect sizes and statistical tests

Calculate effect sizes and conduct basic statistical tests using effect_counts() and effect_metrics(). Effect calculation is included in the reports if you request it by the effect-parameter, for example:

report_counts(data, adopter, sd_gender, prop = "cols", effect = TRUE)

A word of warning: Statistics is the world of uncertainty. All procedures require mindful interpretation. Counting stars might evoke illusions.

Factors and Clusters

You can generate tables and plots for clustering and factor analysis of metric variables. Both clustering and factor analysis are included in the reports when requested using the factors or clusters parameters.

Set the respective parameters to TRUE to generate a scree plot and let the diagnostics choose the optimal number:


report_metrics(data, starts_with("cg_adoption"), factors = TRUE, clusters = TRUE)

Set the desired number directly:


report_metrics(data, starts_with("cg_adoption"), factors = 3, clusters = 4)

You don’t need to add both parameters at the same time if you are only interested in factors or clusters.

Modeling: Regression and Analysis of Variance

Modeling in the statistical sense is predicting an outcome (dependent variable) from one or multiple predictors (independent variables).

The report_metrics() function calculates a linear model if the model parameter is set to TRUE. You provide the variables in the following parameters:

  • Dependent metric variable: first parameter (a single column).
  • Independent categorical variables: second parameter (a tidy column selection).
  • Independent metric variables: metric parameter (a tidy column selection).
  • Interaction effects: interactions parameter with a vector of multiplication terms (e.g. c(sd_age * sd_gender)).
ds |>
 filter(sd_gender != "diverse") |>
 report_metrics(
   use_work, 
   cross = c(sd_gender, adopter), 
   metric = sd_age,
   model = TRUE, diagnostics = TRUE
 )

Four selected diagnostic plots are generated if the diagnostics parameter is set to TRUE:

  • Residual vs. fitted: Residuals should be evenly distributed vertically. Horizontally, they should follow the straight line. Otherwise this could be an indicator for heteroscedasticity, non-linearity, or autocorrelation.
  • Scale-location plot: Points should be evenly distributed without any visible pattern, similar to the residual plot.
  • Q-Q-Plot of fitted values: All points should lie along a straight line. Deviations may suggest non-linearity or that residuals are not normally distributed.
  • Cook’s distance plot: High values indicate that single cases influence the model disproportionately. Rule of thumb: Cook’s distance > 1 is a problem.

To work with the predicted values, use add_model() instead of the report function. This will add a new variable prefixes with prd_, holding the target scores.

ds <- ds |> add_model(
   use_work,
   categorical = c(sd_gender, adopter), 
   metric = sd_age
 )

report_metrics(ds, use_work, prd_use_work, metric = T)

There are two functions to get the regression table or plot from the new column:


model_tab(ds, prd_use_work)
model_plot(ds, prd_use_work)

By default, p-values are adjusted to the number of tests by controlling the false discovery rate (fdr). Set the adjust parameter to FALSE to disable p-value correction.

Reliability Scores (and Classification Performance Indicators)

In content analysis, reliability is usually checked by coding the cases with different persons and then calculating the overlap. To calculate reliability scores, prepare one data frame for each person:

  • All column names in the different data frames should be identical.
    Codings must be either binary (TRUE/FALSE) or contain a fixed number of values such as “sports”, “politics”, “weather”.
  • Add a column holding initials or the name of the coder.
  • One column must contain unique IDs for each case (e.g., a running case number).

Next, you row-bind the data frames. The coder and ID columns ensure that each coding can be uniquely matched to both the coder and the case.


data_coded <- bind_rows(
  data_coder1,
  data_coder2
)

The final data, for example, looks like:

casecodertopic_sportstopic_weather
1anneTRUEFALSE
2anneTRUEFALSE
3anneFALSETRUE
1benTRUETRUE
2benTRUEFALSE
3benFALSETRUE

Calculating reliability is straightforward with report_counts():

  • Provide the data to the first parameter.
  • Add the column with codings (or a selection of multiple columns, e.g. using starts_with()) to the second parameter.
  • Set the the third parameter to the column name with coder names or initials.
  • Set the ids parameter to the column that contains case IDs or case numbers (this tells the volker-package which cases belong together).
  • Set the agree parameter to “reliability” to request reliability scores.

Example:


report_counts(data_coded, starts_with("topic_"), coder, ids = case, prop = "cols", agree = "reliability")

If you are only interested in the scores (without a plot), use agree_tab. Tip: You may abbreviate the method name (e.g., “reli” instead of “reliability”).


agree_tab(data_coded, starts_with("topic_"), coder, ids = case, method = "reli")

You can also request classification performance indicators (accuracy, precision, recall, F1) with the same function by setting the method parameter to “classification” (may be abbreviated). Use this option when comparing manual codings with automated codings (e.g., classifiers or large language models). By default, you get macro statistics (average precision, recall and f1 across categories).

If you have multiple values in a column, you can focus on one category to get micro statistics:


agree_tab(data_coded, starts_with("topic_"), coder, ids = case, method = "class", category = "catcontent")

Where do all the labels go?

One of the strongest package features is labeling. You know the pain. Labels are stored in the column attributes. Inspect current labels of columns and values by the codebook()-function:

codebook(data)

This results in a table with item names, item values, value names and value labels.

You can set specific column labels by providing a named list to the items-parameter of labs_apply():

data %>%
  labs_apply(
    items = list(
      "cg_adoption_advantage_01" = "General advantages",
      "cg_adoption_advantage_02" = "Financial advantages",
      "cg_adoption_advantage_03" = "Work-related advantages",
      "cg_adoption_advantage_04" = "More fun"
    )
  ) %>% 
  tab_metrics(starts_with("cg_adoption_advantage_"))

Labels for values inside a column can be adjusted by providing a named list to the values-parameter of labs_apply(). In addition, select the columns where value labels should be changed:


data %>%
  labs_apply(
    cols = starts_with("cg_adoption"),  
    values = list(
      "1" = "Strongly disagree",
      "2" = "Disagree",
      "3" = "Neutral",
      "4" = "Agree",
      "5" = "Strongly agree"
    ) 
  ) %>% 
  plot_metrics(starts_with("cg_adoption"))

To conveniently manage all labels of a dataset, save the result of codebook() to an Excel file, change the labels manually in a copy of the Excel file, and finally call labs_apply() with your revised codebook.


library(readxl)
library(writexl)

# Save codebook to a file
codes <- codebook(data)
write_xlsx(codes,"codebook.xlsx")

# Load and apply a codebook from a file
codes <- read_xlsx("codebook_revised.xlsx")
data <- labs_apply(data, codes)

Be aware that some data operations such as mutate() from the tidyverse loose labels on their way. In this case, store the labels (in the codebook attribute of the data frame) before the operation and restore them afterwards:

data %>%
  labs_store() %>%
  mutate(sd_age = 2024 - sd_age) %>% 
  labs_restore() %>% 
  
  tab_metrics(sd_age)

SoSci Survey integration

The labeling mechanisms follow a technique used, for example, on SoSci Survey. Sidenote for techies: Labels are stored in the column attributes. That’s why you can directly throw in labeled data from the SoSci Survey API:

library(volker)

# Get your API link from SoSci Survey with settings "Daten als CSV für R abrufen"
eval(parse("https://www.soscisurvey.de/YOURPROJECT/?act=YOURKEY&rScript", encoding="UTF-8"))

# Generate reports
report_counts(ds, A002)

For best results, use sensible prefixes and captions for your SoSci questions. The labels come directly from your questionnaire.

Please note: The values -9, -2, -1 and [NA] nicht beantwortet, [NA] keine Angabe, [no answer] are automatically recoded to missing values within all plot, tab, effect, and report functions. See the clean-parameter help how to disable automatic residual removal.

Customization

You can change plot colors using the theme_vlkr()-function:

theme_set(
  theme_vlkr(
    base_fill = c("#F0983A","#3ABEF0","#95EF39","#E35FF5","#7A9B59"),
    base_gradient = c("#FAE2C4","#F0983A")
  )
)

Plot and table functions share a number of parameters that can be used to customize the outputs. Lookup the available parameters in the help of the specific function.

Data preparation

  • ordered: Sometimes categories have an order, from low to high or from few to many. It helps visual inspections to plot ordered values with shaded colors instead of arbitrary colors. For frequency plots, you can inform the method about the desired order. By default the functions try to automatically detect a sensitive order.
  • category: When you have multiple categories in a column, you can focus one of the categories to simplify the plots and tables. By default, if a column has only TRUE and FALSE values, the outputs focus the TRUE category.
  • clean Before all calculations, the dataset goes through a cleaning plan that, for example, recodes residual factor values such as “[NA] nicht beantwortet” to missings. In surveys, negative values such as -9 or -2 are often used to mark missing values or residual answers (“I don’t know”). See the help for further details or disable data cleaning if you don’t like it. For example, to disable removing the negative residual values, call options(vlkr.na.numbers=FALSE) and options(vlkr.na.levels=FALSE). Rows with missing values in the analyzed columns are, by default, removed before producing tables or plots. To prevent this and to use the maximum available information, call options(vlkr.na.omit=FALSE). From this point on, for example, item summaries are calculated using all values available for each single item. The outputs contain hints about cases with missing data.

Calculations

  • prop: Calculating percentages in a cross tab requires careful selection of the base. You can choose between total, row or column percentages. For stacked bar charts, displaying row percentages instead of total percentages gives a direct visual comparison of groups.
  • ci: Add confidence intervals to plot and table outputs.
  • index: Indexes (=mean of multiple items) can be added to a dataset using add_index() or, using the index-parameter, automatically be included in report functions. Cronbach’s alpha is added to all table outputs.
  • factors: Items can be condensed to factors using principal component analysis. With add_factors() new columns are added. The factors-parameter in the report_metrics function automatically adds a factor analysis to the outputs.
  • clusters: Cases can be grouped to clustersusing kmeans cluster analysis. With add_clusters() a new column indicating cluster groups is added. The clusters-parameter in the report_metrics function automatically adds a cluster analysis to the outputs.
  • effect: You are not sure whether the differences are statistical significant? One option is to look out for non overlapping confidence intervals. In addition, the effect option calculates effect sizes such as Cramer’s v or R squared and generates typical statistical tests such as Chi-squared tests and t-tests.
  • method: By default, correlations are calculated using Pearson’s R. You can choose Spearman’s Rho with the methods-parameter. For cooccurrence analysis of categories (you provide two variable sets to report_counts()), you can choose between Cramer’s V (items are compared with items) or NPMI (the single item values are compared). In this case, we recommend to also set the tiles-parameter to TRUE for generating a heatmap (the default are bar plots). Normalized pointwise mutual information helps spotting combinations that are more rare (negative values) or more frequent (positive values) than expected by chance. You can add “npmi” to the numbers parameter to plot the values on the tiles.
  • adjust: Performing multiple significance tests inflates the alpha error. Thus, p values need to be adjusted according to the number of tests. By default, the false discovery rate is controlled (fdr). You can, for example, choose Bonferroni-correction or disable correction.

Labeling

  • title: All plots usually get a title derived from the column attributes or column names. Set to FALSE to suppress the title or provide a title of your choice as a character value.
  • labels: Labels are extracted from the column attributes, if present. Set to FALSE to output bare column names and values.

Tables

  • percent: Frequency tables show percentages by default. Set to FALSE to get raw proportions - easier to postprocess in further calculations.
  • digits: Tables containing means and standard deviations by default round values to one digit. Increase the number to show more digits.
  • values: The more variables you desire, the denser the output must be. Some tables try to serve you insights at the maximum and show two values in one cell, for example the absolute counts (n) and the percentages (p), or the mean (m) and the standard deviation (sd). Control your desire with the values-parameter.

Plots

  • numbers: Bar plots give quick impressions, tables provide exact numbers. In bar charts you can combine both and print the frequencies onto the bars. Set the numbers parameter to “n”, “p” or c(“n”,“p”). To prevent cluttering and overlaps, numbers are only plotted on bars larger than 5%.
  • limits: Do you know how to create misleading graphs? It happens when you truncate the minimum or maximum value in a scale. The scale limits are automatically guessed by the package functions (work in progress). Use the limits-parameter to manually fix any misleading graphs.
  • box: In metric plots you can visualise the distribution by adding boxplots.
  • log: Metric values having long tail distributions are not easy to visualise. In scatter plots, you can use a logarithmic scale. Be aware, that zero values will be omitted because their log value is undefined.
  • width: In stacked plots with column or row proportions, by default, the column or bar widths mirror the group size. Set to FALSE to produce equal widths.
  • tiles: The default plot for two categorical variables is a bar plot. Alternatively, generate a heat map.

Installation

As with all other packages you’ll have to install the package first.

install.packages("strohne/volker")

You can try alternative versions:

  • If you want, install the main version from GitHub using remotes, which may include features not yet published on CRAN (if asked, skip the updates):

    if (!require(remotes)) { install.packages("remotes") }
    remotes::install_github("strohne/volker", upgrade="never", build_vignettes = TRUE)
  • In case you are adventurous, try the latest experimental development version which lives in the devel branch (if asked, skip the updates):

    if (!require(remotes)) { install.packages("remotes") }
    remotes::install_github("strohne/volker", ref="devel", upgrade="never", build_vignettes = TRUE)

Special features

  • Simple tables, simple plots, simple reports.
  • Labeling and scaling based on attributes. Appropriate attributes, for example, are provided by the SoSci Survey API. Alternatively, you can add custom labels. Use codebook() to see all labels present in a dataset.
  • Interactive reports: Use the volker::html_report template in your Markdown documents to switch between tables and plots when using the report-functions.
  • Generate metric indexes, conduct simple factor and cluster analyses and calculate effect sizes
  • Simplified hints for wrong parameters, e.g. if you forget to provide a data frame (work in progress).
  • Tidyverse style.

Troubleshooting

The kableExtra package produces an error in R 4.3 when knitting documents: .onLoad in loadNamespace() für 'kableExtra' fehlgeschlagen. As a work around, remove PDF and Word settings from the output options in you markdown document (the yml section at the top). Alternatively, install the latest development version:

remotes::install_github("kupietz/kableExtra")

Roadmap

VersionFeaturesStatus
1.0Descriptives90% done
2.0Effects75% done
3.0Factors & clusters80% done
4.0Linear models50% done
5.0Text analysiswork in progress

Similar packages

The volker package is inspired by outputs used in the textbook Einfache Datenauswertung mit R (Gehrau & Maubach et al., 2022), which provides an introduction to univariate and bivariate statistics and data representation using RStudio and R Markdown.

Other packages with high-level reporting functions:

Authors and citation

Authors
Jakob Jünger (University of Münster)
Henrieke Kotthoff (University of Münster)

Contributers
Chantal Gärtner (University of Münster)

Citation
Jünger, J. & Kotthoff, H. (2024). volker: High-level functions for tabulating, charting and reporting survey data. R package version 3.0.

Copy Link

Version

Install

install.packages('volker')

Monthly Downloads

233

Version

3.2.0

License

MIT + file LICENSE

Issues

Pull Requests

Stars

Forks

Maintainer

Jakob Jünger

Last Published

October 10th, 2025

Functions in volker (3.2.0)

cluster_plot

Get plot for clustering result
codebook

Get variable and value labels from a data set
data_cat

Convert numeric values to string
cfg_get_na_numbers

Get configured na numbers
VLKR_NA_OMIT

Whether to remove all missings
VLKR_NA_NUMBERS

Numbers to remove from vectors
data_clean

Prepare dataframe for the analysis
VLKR_FILLGRADIENT

Gradient colors
chatgpt

ChatGPT Adoption Dataset CG-GE-APR23
add_clusters

Add cluster number to a data frame
add_factors

Add PCA columns along with summary statistics (KMO and Bartlett test) to a data frame
data_num

Convert values to numeric values
VLKR_FILLPOLARIZED

Polarized colors
check_is_dataframe

Check whether the object is a dataframe
VLKR_FIG_SETTINGS

Resolution settings for plots
data_clean_default

Prepare data originating from SoSci Survey or SPSS
VLKR_FILLDISCRETE

Fill colors
data_onehot

One-hot encode selected columns
data_round

Round and format selected numeric columns
data_rm_zeros

Remove zero values, drop missings and output a message
add_index

Calculate the mean value of multiple items
VLKR_POINT_ALPHA

Alpha values
add_model

Add a column with predicted values from a regression model
check_nonequal_columns

Check whether there is no overlap between two column sets
check_is_param

Check whether a parameter value is from a valid set
check_is_numeric

Check whether a column selection is numeric
VLKR_POINT_MEAN_SHAPE

Shapes
check_is_categorical

Check whether a column selection is categorical
check_has_column

Check whether a column exist and stop if not
diagnostics_qq

Normal Q-Q
data_rev

Reverse item values
diagnostics_cooksd

Cook's distance plot
data_prepare

Prepare data for calculation
.plot_bars

Helper function: plot grouped bar chart
data_rm_missings

Remove missings and output a message
data_rm_na_numbers

Remove NA numbers
data_rm_negatives

Remove negatives and output a warning
.plot_heatmap

Helper function: Heatmap
diagnostics_resid_fitted

Residuals vs Fitted plot
.pair_disagreement

Helper: mean of pairwise disagreements (nominal)
.plot_cor

Helper function: plot cor and regression outputs
.attr_setcolumn

Set an attribute value on selected columns of a data frame
.agree_reliability

Calculate reliability scores
diagnostics_scale_location

Scale-Location (Spread-Location)
.tab_split

Split a metric column into categories based on the median
data_rm_na_levels

Remove NA levels
.attr_insert

Insert a name-value-pair into an object attribute
.report_idx

Generate an index table and plot
.knit_prepare

Prepare markdown content for table rendering
.knit_plot

Knit volker plots
.attr_transfer

Transfer attributes from one to another object
effect_counts_one

Test homogeneity of category shares
.to_vlkr_df

Add vlkr_df class - that means, the data frame has been prepared
effect_counts_items_grouped

Effect size and test for comparing multiple variables by a grouping variable
.report_mdl

Generate an model table and plot
.knit_shorten

Compact table printing with shortened names and values
.whisker_upper

Calculate upper whisker in a boxplot
.whisker_lower

Calculate lower whisker in a boxplot
.knit_table

Knit volker tables
effect_counts_one_cor

Output test statistics and effect size from a logistic regression of one metric predictor
.plot_lines

Helper function: plot grouped line chart
effect_counts_items_cor

Correlate the values in multiple items with one metric column and output effect sizes and tests
.agree_confusion

Generate confusion matrix
.agree_items

Calculate agreement coefficients for multiple items
effect_counts_items_grouped_items

Effect size and test for comparing multiple variables by multiple grouping variables
.factor_with_attr

Create a factor vector and preserve all attributes
.effect_correlations

Calculate correlation and cooccurrence coefficients and test whether they are different from zero
effect_metrics_items_cor_items

Output correlation coefficients for multiple items
effect_metrics_items_grouped

Compare groups for each item by calculating F-statistics and effect sizes
effect_counts_items_cor_items

Correlate the values in multiple items with multiple metric columns and output effect sizes and tests
label_scale

Wrap labels in plot scales
model_metrics_tab

Output a regression table with estimates and macro statistics for multiple categorical or metric independent variables
labs_apply

Set column and value labels
plot_metrics

Output a plot with distribution parameters such as the mean values
mutate

Mutate function
plot_counts_items_grouped

Plot percent shares of multiple items compared by groups
.plot_scree

Helper function: scree plot
.outliers

Calculate outliers
ends_with

Select variables by their postfix
factor_plot

Get plot with factor analysis result
plot_counts_items_cor_items

Correlation of categorical items with metric items
get_direction

Detect whether a scale is a numeric sequence
plot_metrics_one_grouped

Output averages for multiple variables
plot_metrics_items

Output averages for multiple variables
.to_vlkr_tab

Add vlkr_tbl class
.to_vlkr_rprt

Add the vlkr_rprt class to an object
.report_fct

Generate an factor table and plot
.pair_agreement

Helper: mean of pairwise agreements (nominal)
get_alpha

Get number of items and Cronbach's alpha of a scale added by add_index()
.report_cls

Generate an cluster table and plot
.to_vlkr_list

Add vlkr_list class
effect_counts_one_grouped

Output test statistics and effect size for contingency tables
tab_counts_items

Output frequencies for multiple variables
prepare_scale

Prepare the scale attribute values
.add_to_vlkr_rprt

Add an object to the report list
effect_metrics

Output effect sizes and test statistics for metric data
factor_tab

Get tables with factor analysis results
tab_metrics_one_grouped

Output a five point summary for groups
theme_set

Get, set, and modify the active ggplot theme
tab_counts_items_cor

Compare the values in multiple items by a metric column that will be split into groups
.agree_classification

Calculate classification performance indicators such as precision and recall.
.iqr

Calculate IQR
.get_plot_limits

Guess plot limits from values
.get_density_mode

Get the maximum density value in a density plot
.plot_summary

Helper function: plot grouped line chart by summarising values
.get_fig_settings

Get plot size and resolution for the current output format from the config
get_ci

Calculate ci values to be used for error bars on a plot
get_correlation

Calculate correlation between two vectors
filter

Filter function
get_prefix

Get the common prefix of character values
labs_clear

Remove all comments from the selected columns
labs_impute

Add missing residual labels in numeric columns that have at least one labeled value
.to_vlkr_plot

Add the volker class and options
get_angle

Angle labels
get_etasq

Calculate Eta squared
get_limits

Get the numeric range from the labels
get_stars

Get significance stars from p values
%>%

Pipe operator
effect_metrics_items

Test whether a distribution is normal for each item
effect_metrics_items_cor

Output correlation coefficients for items and one metric variable
effect_counts

Output effect sizes and test statistics for count data
.report_agr

Generate table and plot for agreement analysis
plot_counts

Output a frequency plot
plot_metrics_items_cor_items

Heatmap for correlations between multiple items
print.vlkr_rprt

Printing method for volker reports
plot_metrics_items_cor

Multiple items correlated with one metric variable
print.vlkr_tbl

Printing method for volker tables.
tab_counts_items_grouped_items

Correlation of categorical items with categorical items
tab_metrics_items_grouped

Output the means for groups in one or multiple columns
tab_metrics_items_grouped_items

Correlation of metric items with categorical items
tab_counts_one

Output a frequency table for the values in one column
trim_label

Remove trailing zeros and trailing or leading whitespaces, colons, hyphens and underscores
trim_prefix

Remove a prefix from a character vector or a factor
wrap_label

Wrap a string
pdf_report

Volker style PDF document format
named.to.list

Convert a named vector to a list
volker-package

volker: High-Level Functions for Tabulating, Charting and Reporting Survey Data
effect_metrics_one_cor

Test whether the correlation is different from zero
plot_counts_items_cor

Plot percent shares of multiple items compared by a metric variable split into groups
plot_counts_one_grouped

Plot frequencies cross tabulated with a grouping column
model_metrics_plot

Plot regression coefficients
plot_counts_one_cor

Plot frequencies cross tabulated with a metric column that will be split into groups
effect_metrics_one_grouped

Output a regression table with estimates and macro statistics
plot_counts_items

Output frequencies for multiple variables
map_label

Get a label for a key
get_baseline

Get a formatted baseline from attributes of an object.
get_npmi

Calculate nmpi
get_betas

Calculate standardized betas
skim_grouped

Calculate a metric by groups
skim_boxplot

A skimmer for boxplot generation
plot_metrics_items_grouped

Output averages for multiple variables compared by a grouping variable
skim_metrics

A reduced skimmer for metric variables Returns a five point summary, mean and sd, items count and alpha for scales added by add_index()
select

Select function
tab_metrics_items_cor

Output a correlation table for item battery and one metric variable
plot_metrics_items_grouped_items

Correlation of metric items with categorical items
theme_vlkr

Define a default theme for volker plots
print.vlkr_plt

Printing method for volker plots
tab_metrics_one

Output a five point summary table for the values in multiple columns
print.vlkr_list

Printing method for volker lists
tibble

Tidy tibbles
tab_metrics_items_cor_items

Output a correlation table for item battery and item battery
tab_metrics_one_cor

Correlate two columns
html_report

Volker style HTML document format
get_title

Get a common title for a column selection
labs_restore

Restore labels from the codebook store in the codebook attribute.
labs_store

Get the current codebook and store it in the codebook attribute.
effect_metrics_one

Test whether a distribution is normal
effect_counts_items

Test homogeneity of category shares for multiple items
plot_metrics_one

Output a density plot for a single metric variable
effect_metrics_items_grouped_items

Compare groups for each item with multiple target items by calculating F-statistics and effect sizes
plot_metrics_one_cor

Correlate two items
tidy_lm_levels

Tidy lm results, replace categorical parameter names by their levels and add the reference level
zip_tables

Combine two identically shaped data frames by adding values of each column from the second data frame into the corresponding column in the first dataframe using parentheses
get_labels

Get the labels of values from a codebook
knit_print.vlkr_plt

Printing method for volker plots when knitting
tribble

Tidy tribbles
get_gini

Calculate the Gini coefficient
idx_add

Deprecated Alias for add_index
report_counts

Create table and plot for categorical variables
labs_prefix

Remove common prefix from the first column
tab_counts

Output a frequency table
report_metrics

Create table and plot for metric variables
starts_with

Select variables by their prefix
plot_counts_items_grouped_items

Correlation of categorical items with categorical items
labs_replace

Replace item value names in a column by their labels
tab_counts_one_cor

Count values by a metric column that will be split into groups
plot_counts_one

Plot the frequency of values in one column
tab_counts_items_cor_items

Correlation of categorical items with metric items
vlkr_colors_discrete

Get colors for discrete scales
vlkr_colors_contrast

Given a vector of hex fill colors, choose an appropriate color for text
tab_counts_one_grouped

Output frequencies cross tabulated with a grouping column
tab_metrics

Output a table with distribution parameters
tab_metrics_items

Output a five point summary table for multiple items
trunc_labels

Truncate labels
tab_counts_items_grouped

Compare the values in multiple items by a grouping column
vlkr_alpha_interpolated

Interpolate an alpha value based on case numbers
vlkr_colors_polarized

Get colors for polarized scales
vlkr_colors_sequential

Get colors for sequential scales
VLKR_NA_LEVELS

Levels to remove from factors
VLKR_NORMAL_DIGITS

Output thresholds
VLKR_MAX_CATEGORIES

Maximum number of distinct values to determine whether a column selection contains only categorical values
VLKR_POINT_SIZE

Sizes
adjust_p

Adjust p-values from multiple tests and optionally annotate significance stars
agree_tab

Agreement for multiple items
VLKR_PLOT_LABELWRAP

Wrapping threshold
VLKR_WRAP_SEPARATOR

Word wrap separators
cluster_tab

Get tables for clustering result