Last chance! 50% off unlimited learning
Sale ends in
Calculation of an overall proportion from studies reporting a
single proportion. Inverse variance method and generalised linear
mixed model (GLMM) are available for pooling. For GLMMs, the
rma.glmm
function from R package
metafor (Viechtbauer 2010) is called internally.
metaprop(event, n, studlab, data = NULL, subset = NULL,
exclude = NULL, method, sm = gs("smprop"), incr = gs("incr"),
allincr = gs("allincr"), addincr = gs("addincr"),
method.ci = gs("method.ci"), level = gs("level"),
level.comb = gs("level.comb"), comb.fixed = gs("comb.fixed"),
comb.random = gs("comb.random"), hakn = gs("hakn"), method.tau,
tau.preset = NULL, TE.tau = NULL, tau.common = gs("tau.common"),
prediction = gs("prediction"), level.predict = gs("level.predict"),
null.effect = NA, method.bias = gs("method.bias"),
backtransf = gs("backtransf"), pscale = 1, title = gs("title"),
complab = gs("complab"), outclab = "", byvar, bylab,
print.byvar = gs("print.byvar"), byseparator = gs("byseparator"),
keepdata = gs("keepdata"), warn = gs("warn"), control = NULL, ...)
Number of events.
Number of observations.
An optional vector with study labels.
An optional data frame containing the study information, i.e., event and n.
An optional vector specifying a subset of studies to be used.
An optional vector specifying studies to exclude from meta-analysis, however, to include in printouts and forest plots.
A character string indicating which method is to be
used for pooling of studies. One of "Inverse"
and
"GLMM"
, can be abbreviated.
A character string indicating which summary measure
("PFT"
, "PAS"
, "PRAW"
, "PLN"
, or
"PLOGIT"
) is to be used for pooling of studies, see
Details.
A numeric which is added to event number and sample size of studies with zero or all events, i.e., studies with an event probability of either 0 or 1.
A logical indicating if incr
is considered
for all studies if at least one study has either zero or all
events. If FALSE (default), incr
is considered only in
studies with zero or all events.
A logical indicating if incr
is used for all
studies irrespective of number of events.
A character string indicating which method is used to calculate confidence intervals for individual studies, see Details.
The level used to calculate confidence intervals for individual studies.
The level used to calculate confidence intervals for pooled estimates.
A logical indicating whether a fixed effect meta-analysis should be conducted.
A logical indicating whether a random effects meta-analysis should be conducted.
A logical indicating whether the method by Hartung and Knapp should be used to adjust test statistics and confidence intervals.
A character string indicating which method is
used to estimate the between-study variance
Prespecified value for the square-root of the
between-study variance
Overall treatment effect used to estimate the between-study variance tau-squared.
A logical indicating whether tau-squared should be the same across subgroups.
A logical indicating whether a prediction interval should be printed.
The level used to calculate prediction interval for a new study.
A numeric value specifying the effect under the null hypothesis.
A character string indicating which test is to
be used. Either "rank"
, "linreg"
, or "mm"
,
can be abbreviated. See function metabias
.
A logical indicating whether results for
transformed proportions (argument sm != "PRAW"
) should be
back transformed in printouts and plots. If TRUE (default),
results will be presented as proportions; otherwise transformed
proportions will be shown. See Details for presentation of
confidence intervals.
A numeric defining a scaling factor for printing of single event probabilities.
Title of meta-analysis / systematic review.
Comparison label.
Outcome label.
An optional vector containing grouping information
(must be of same length as event
).
A character string with a label for the grouping variable.
A logical indicating whether the name of the grouping variable should be printed in front of the group labels.
A character string defining the separator between label and levels of grouping variable.
A logical indicating whether original data (set) should be kept in meta object.
A logical indicating whether the addition of
incr
to studies with zero or all events should result in a
warning.
Additional arguments passed on to
rma.glmm
function.
An object of class c("metaprop", "meta")
with corresponding
print
, summary
, and forest
functions. The
object is a list containing the following components:
As defined above.
As defined above.
As defined above.
As defined above.
As defined above.
As defined above.
As defined above.
Estimated (un)transformed proportion and its standard error for individual studies.
Lower and upper confidence interval limits for individual studies.
z-value and p-value for test of treatment effect for individual studies.
Weight of individual studies (in fixed and random effects model).
Estimated overall (un)transformed proportion and standard error (fixed effect model).
Lower and upper confidence interval limits (fixed effect model).
z-value and p-value for test of overall effect (fixed effect model).
Estimated overall (un)transformed proportion and standard error (random effects model).
Lower and upper confidence interval limits (random effects model).
z-value or t-value and corresponding p-value for test of overall effect (random effects model).
As defined above.
Standard error utilised for prediction interval.
Lower and upper limits of prediction interval.
Number of studies combined in meta-analysis.
Heterogeneity statistic Q.
Degrees of freedom for heterogeneity statistic.
P-value of heterogeneity test.
Heterogeneity statistic for likelihood-ratio test
(only if method = "GLMM"
).
Degrees of freedom for likelihood-ratio test
P-value of likelihood-ratio test.
Square-root of between-study variance.
Standard error of square-root of between-study variance.
Scaling factor utilised internally to calculate common tau-squared across subgroups.
A character string indicating method used for
pooling: "Inverse"
Degrees of freedom for test of treatment effect for
Hartung-Knapp method (only if hakn=TRUE
).
Levels of grouping variable - if byvar
is not
missing.
Estimated treatment effect and
standard error in subgroups (fixed effect model) - if
byvar
is not missing.
Lower and upper confidence
interval limits in subgroups (fixed effect model) - if
byvar
is not missing.
z-value and p-value for test of
treatment effect in subgroups (fixed effect model) - if
byvar
is not missing.
Estimated treatment effect and
standard error in subgroups (random effects model) - if
byvar
is not missing.
Lower and upper confidence
interval limits in subgroups (random effects model) - if
byvar
is not missing.
z-value or t-value and
corresponding p-value for test of treatment effect in subgroups
(random effects model) - if byvar
is not missing.
Weight of subgroups (in fixed and
random effects model) - if byvar
is not missing.
Degrees of freedom for test of treatment effect
for Hartung-Knapp method in subgroups - if byvar
is not
missing and hakn=TRUE
.
Harmonic mean of number of observations in
subgroups (for back transformation of Freeman-Tukey Double
arcsine transformation) - if byvar
is not missing.
Number of events in subgroups - if byvar
is
not missing.
Number of observations in subgroups - if byvar
is
not missing.
Number of studies combined within subgroups - if
byvar
is not missing.
Number of all studies in subgroups - if byvar
is not missing.
Overall within subgroups heterogeneity statistic Q
(based on fixed effect model) - if byvar
is not missing.
Overall within subgroups heterogeneity statistic
Q (based on random effects model) - if byvar
is not
missing (only calculated if argument tau.common
is TRUE).
Degrees of freedom for test of overall within
subgroups heterogeneity - if byvar
is not missing.
P-value of within subgroups heterogeneity
statistic Q (based on fixed effect model) - if byvar
is
not missing.
P-value of within subgroups heterogeneity
statistic Q (based on random effects model) - if byvar
is
not missing.
Overall between subgroups heterogeneity statistic
Q (based on fixed effect model) - if byvar
is not
missing.
Overall between subgroups heterogeneity statistic
Q (based on random effects model) - if byvar
is not
missing.
Degrees of freedom for test of overall between
subgroups heterogeneity - if byvar
is not missing.
P-value of between subgroups heterogeneity
statistic Q (based on fixed effect model) - if byvar
is
not missing.
P-value of between subgroups heterogeneity
statistic Q (based on random effects model) - if byvar
is
not missing.
Square-root of between-study variance within subgroups
- if byvar
is not missing.
Scaling factor utilised internally to calculate common
tau-squared across subgroups - if byvar
is not missing.
Heterogeneity statistic H within subgroups - if
byvar
is not missing.
Lower and upper confidence limti for
heterogeneity statistic H within subgroups - if byvar
is
not missing.
Heterogeneity statistic I2 within subgroups - if
byvar
is not missing.
Lower and upper confidence limti for
heterogeneity statistic I2 within subgroups - if byvar
is
not missing.
Increment added to number of events.
As defined above.
Original data (set) used in function call (if
keepdata=TRUE
).
Information on subset of original data used in
meta-analysis (if keepdata=TRUE
).
GLMM object generated by call of
rma.glmm
function (fixed effect model).
GLMM object generated by call of
rma.glmm
function (random effects model).
Function call.
Version of R package meta used to create object.
Version of R package metafor used for GLMMs.
This function provides methods for fixed effect and random effects
meta-analysis of single proportions to calculate an overall
proportion. Note, you should use R function metabin
to compare proportions of pairwise comparisons instead of using
metaprop
for each treatment arm separately which will break
randomisation in randomised controlled trials.
The following transformations of proportions are implemented to calculate an overall proportion:
Logit transformation (sm = "PLOGIT"
, default)
Log transformation (sm = "PLN"
)
Freeman-Tukey Double arcsine transformation (sm = "PFT"
)
Arcsine transformation (sm = "PAS"
)
Raw, i.e. untransformed, proportions (sm = "PRAW"
)
Classic meta-analysis (Borenstein et al., 2010) utilises the
transformed proportions and corresponding standard errors in the
generic inverse variance method. A distinctive and frequently
overlooked advantage of binary data is that individual patient data
can be extracted. Accordingly, a generalised linear mixed model
(GLMM) - more specific, a random intercept logistic regression
model - can be utilised for the meta-analysis of proportions
(Stijnen et al., 2010). This method - implicitly using the logit
transformation - is available (argument method = "GLMM"
) by
calling the rma.glmm
function from R package
metafor internally.
For the logit transformation, a random intercept logistic
regression model is used by default, i.e., argument method =
"GLMM"
. The classic meta-analysis model based on the inverse
variance method can be used instead by setting argument
method
equal to "Inverse"
.
Contradictory recommendations on the use of transformations of proportions have been published in the literature. For example, Barendregt et al. (2013) recommend the use of the Freeman-Tukey double arcsine transformation instead of the logit transformation whereas Warton & Hui (2011) strongly advise to use generalised linear mixed models with the logit transformation instead of the arcsine transformation. Schwarzer et al. (2019) describe seriously misleading results in a meta-analysis with very different sample sizes due to problems with the back-transformation of the Freeman-Tukey transformation which requires a single sample size. Accordingly, Schwarzer et al. (2019) also recommend to use GLMMs for the meta-analysis of single proportions, however, admit that individual study weights are not available with this method. Meta-analysts which require individual study weights should consider the arcsine or logit transformation.
In order to prevent misleading conclusions for the Freeman-Tukey double arcsine transformation, sensitivity analyses using other transformations or using a range of sample sizes should be conducted (Schwarzer et al., 2019).
Various methods are available to calculate confidence intervals for individual study results (see Agresti & Coull 1998 and Newcombe 1988):
Clopper-Pearson interval also called 'exact' binomial
interval (method.ci = "CP"
, default)
Wilson Score interval (method.ci = "WS"
)
Wilson Score interval with continuity correction
(method.ci = "WSCC"
)
Agresti-Coull interval (method.ci = "AC"
)
Simple approximation interval (method.ci = "SA"
)
Simple approximation interval with continuity correction
(method.ci = "SACC"
)
Normal approximation interval based on summary measure,
i.e. defined by argument sm
(method.ci = "NAsm"
)
Note, with exception of the normal approximation based on the
summary measure, i.e. method.ci = "NAsm"
, the same
confidence interval is calculated for individual studies for any
summary measure (argument sm
) as only number of events and
observations are used in the calculation disregarding the chosen
summary measure. Results will be presented for transformed
proportions if argument backtransf = FALSE
in the
print.meta
, print.summary.meta
, or
forest.meta
function. In this case, argument
method.ci = "NAsm"
is used, i.e. confidence intervals based
on the normal approximation based on the summary measure.
Argument pscale
can be used to rescale proportions, e.g.
pscale = 1000
means that proportions are expressed as events
per 1000 observations. This is useful in situations with (very) low
event probabilities.
For several arguments defaults settings are utilised (assignments
using gs
function). These defaults can be changed
using the settings.meta
function.
Internally, both fixed effect and random effects models are
calculated regardless of values choosen for arguments
comb.fixed
and comb.random
. Accordingly, the estimate
for the random effects model can be extracted from component
TE.random
of an object of class "meta"
even if
argument comb.random = FALSE
. However, all functions in R
package meta will adequately consider the values for
comb.fixed
and comb.random
. E.g. function
print.meta
will not print results for the random
effects model if comb.random = FALSE
.
If the summary measure is equal to "PRAW", "PLN", or "PLOGIT", a
continuity correction is applied if any study has either zero or
all events, i.e., an event probability of either 0 or 1. By
default, 0.5 is used as continuity correction (argument
incr
). This continuity correction is used both to calculate
individual study results with confidence limits and to conduct
meta-analysis based on the inverse variance method. For GLMMs no
continuity correction is used.
Argument byvar
can be used to conduct subgroup analysis for
all methods but GLMMs. Instead use the metareg
function for GLMMs which can also be used for continuous
covariates.
A prediction interval for the treatment effect of a new study is
calculated (Higgins et al., 2009) if arguments prediction
and comb.random
are TRUE
.
R function update.meta
can be used to redo the
meta-analysis of an existing metaprop object by only specifying
arguments which should be changed.
For the random effects, the method by Hartung and Knapp (2003) is
used to adjust test statistics and confidence intervals if argument
hakn = TRUE
.
The DerSimonian-Laird estimate (1986) is used in the random effects
model if method.tau = "DL"
. The iterative Paule-Mandel
method (1982) to estimate the between-study variance is used if
argument method.tau = "PM"
. Internally, R function
paulemandel
is called which is based on R function
mpaule.default from R package metRology from S.L.R. Ellison
<s.ellison at lgc.co.uk>.
If R package metafor (Viechtbauer 2010) is installed, the
following methods to estimate the between-study variance
method.tau
) are also available:
Restricted maximum-likelihood estimator (method.tau =
"REML"
)
Maximum-likelihood estimator (method.tau = "ML"
)
Hunter-Schmidt estimator (method.tau = "HS"
)
Sidik-Jonkman estimator (method.tau = "SJ"
)
Hedges estimator (method.tau = "HE"
)
Empirical Bayes estimator (method.tau = "EB"
)
For these methods the R function rma.uni
of R package
metafor is called internally. See help page of R function
rma.uni
for more details on these methods to estimate
between-study variance.
Agresti A & Coull BA (1998): Approximate is better than "exact" for interval estimation of binomial proportions. The American Statistician, 52, 119--26
Barendregt JJ, Doi SA, Lee YY, Norman RE, Vos T (2013): Meta-analysis of prevalence. Journal of Epidemiology and Community Health, 67, 974--8
Borenstein M, Hedges LV, Higgins JP, Rothstein HR (2010): A basic introduction to fixed-effect and random-effects models for meta-analysis. Research Synthesis Methods, 1, 97--111
DerSimonian R & Laird N (1986): Meta-analysis in clinical trials. Controlled Clinical Trials, 7, 177--88
Edward JM et al. (2006): Adherence to antiretroviral therapy in sub-saharan Africa and North America - a meta-analysis. Journal of the American Medical Association, 296, 679--90
Freeman MF & Tukey JW (1950): Transformations related to the angular and the square root. Annals of Mathematical Statistics, 21, 607--11
Higgins JPT, Thompson SG, Spiegelhalter DJ (2009): A re-evaluation of random-effects meta-analysis. Journal of the Royal Statistical Society: Series A, 172, 137--59
Knapp G & Hartung J (2003): Improved tests for a random effects meta-regression with a single covariate. Statistics in Medicine, 22, 2693--710
Miller JJ (1978): The inverse of the Freeman-Tukey double arcsine transformation. The American Statistician, 32, 138
Newcombe RG (1998): Two-sided confidence intervals for the single proportion: comparison of seven methods. Statistics in Medicine, 17, 857--72
Paule RC & Mandel J (1982): Consensus values and weighting factors. Journal of Research of the National Bureau of Standards, 87, 377--85
Pettigrew HM, Gart JJ, Thomas DG (1986): The bias and higher cumulants of the logarithm of a binomial variate. Biometrika, 73, 425--35
Schwarzer G, Chemaitelly H, Abu-Raddad LJ, R<U+00FC>cker G (2019): Seriously misleading results using inverse of Freeman-Tukey double arcsine transformation in meta-analysis of single proportions. Research Synthesis Methods, 1--8. https://doi.org/10.1002/jrsm.1348
Stijnen T, Hamza TH, Ozdemir P (2010): Random effects meta-analysis of event outcome in the framework of the generalized linear mixed model with applications in sparse data. Statistics in Medicine, 29, 3046--67
Viechtbauer W (2010): Conducting meta-analyses in R with the metafor package. Journal of Statistical Software, 36, 1--48
Warton DI, Hui FKC (2011): The arcsine is asinine: the analysis of proportions in ecology. Ecology, 92, 3--10
# NOT RUN {
# Meta-analysis using generalised linear mixed model
#
metaprop(4:1, 10 * 1:4)
# Apply various classic meta-analysis methods to estimate
# proportions
#
m1 <- metaprop(4:1, 10 * 1:4, method = "Inverse")
m2 <- update(m1, sm = "PAS")
m3 <- update(m1, sm = "PRAW")
m4 <- update(m1, sm = "PLN")
m5 <- update(m1, sm = "PFT")
#
m1
m2
m3
m4
m5
#
forest(m1)
# }
# NOT RUN {
forest(m2)
forest(m3)
forest(m3, pscale = 100)
forest(m4)
forest(m5)
# }
# NOT RUN {
# Do not back transform results, e.g. print logit transformed
# proportions if sm = "PLOGIT" and store old settings
#
oldset <- settings.meta(backtransf = FALSE)
#
m6 <- metaprop(4:1, c(10, 20, 30, 40), method = "Inverse")
m7 <- update(m6, sm = "PAS")
m8 <- update(m6, sm = "PRAW")
m9 <- update(m6, sm = "PLN")
m10 <- update(m6, sm = "PFT")
#
forest(m6)
# }
# NOT RUN {
forest(m7)
forest(m8)
forest(m8, pscale = 100)
forest(m9)
forest(m10)
# }
# NOT RUN {
# Use old settings
#
settings.meta(oldset)
# Examples with zero events
#
m1 <- metaprop(c(0, 0, 10, 10), rep(100, 4), method = "Inverse")
m2 <- metaprop(c(0, 0, 10, 10), rep(100, 4), incr = 0.1, method = "Inverse")
#
summary(m1)
summary(m2)
#
# }
# NOT RUN {
forest(m1)
forest(m2)
# }
# NOT RUN {
# Example from Miller (1978):
#
death <- c(3, 6, 10, 1)
animals <- c(11, 17, 21, 6)
#
m3 <- metaprop(death, animals, sm = "PFT")
forest(m3)
# Data examples from Newcombe (1998)
# - apply various methods to estimate confidence intervals for
# individual studies
#
event <- c(81, 15, 0, 1)
n <- c(263, 148, 20, 29)
#
m1 <- metaprop(event, n, method.ci = "SA", method = "Inverse")
m2 <- update(m1, method.ci = "SACC")
m3 <- update(m1, method.ci = "WS")
m4 <- update(m1, method.ci = "WSCC")
m5 <- update(m1, method.ci = "CP")
#
lower <- round(rbind(NA, m1$lower, m2$lower, NA, m3$lower,
m4$lower, NA, m5$lower), 4)
upper <- round(rbind(NA, m1$upper, m2$upper, NA, m3$upper,
m4$upper, NA, m5$upper), 4)
#
tab1 <- data.frame(
scen1 = meta:::formatCI(lower[, 1], upper[, 1]),
scen2 = meta:::formatCI(lower[, 2], upper[, 2]),
scen3 = meta:::formatCI(lower[, 3], upper[, 3]),
scen4 = meta:::formatCI(lower[, 4], upper[, 4]),
stringsAsFactors = FALSE
)
names(tab1) <- c("r=81, n=263", "r=15, n=148",
"r=0, n=20", "r=1, n=29")
row.names(tab1) <- c("Simple", "- SA", "- SACC",
"Score", "- WS", "- WSCC",
"Binomial", "- CP")
tab1[is.na(tab1)] <- ""
# Newcombe (1998), Table I, methods 1-5:
tab1
# Same confidence interval, i.e. unaffected by choice of summary
# measure
#
print(metaprop(event, n, method.ci = "WS", method = "Inverse"), ma = FALSE)
print(metaprop(event, n, sm = "PLN", method.ci = "WS"), ma = FALSE)
print(metaprop(event, n, sm = "PFT", method.ci = "WS"), ma = FALSE)
print(metaprop(event, n, sm = "PAS", method.ci = "WS"), ma = FALSE)
print(metaprop(event, n, sm = "PRAW", method.ci = "WS"), ma = FALSE)
# Different confidence intervals as argument sm = "NAsm"
#
print(metaprop(event, n, method.ci = "NAsm", method = "Inverse"), ma = FALSE)
print(metaprop(event, n, sm = "PLN", method.ci = "NAsm"), ma = FALSE)
print(metaprop(event, n, sm = "PFT", method.ci = "NAsm"), ma = FALSE)
print(metaprop(event, n, sm = "PAS", method.ci = "NAsm"), ma = FALSE)
print(metaprop(event, n, sm = "PRAW", method.ci = "NAsm"), ma = FALSE)
# Different confidence intervals as argument backtransf = FALSE.
# Accordingly, method.ci = "NAsm" used internally.
#
print(metaprop(event, n, method.ci = "WS", method = "Inverse"),
ma = FALSE, backtransf = FALSE)
print(metaprop(event, n, sm = "PLN", method.ci = "WS"),
ma = FALSE, backtransf = FALSE)
print(metaprop(event, n, sm = "PFT", method.ci = "WS"),
ma = FALSE, backtransf = FALSE)
print(metaprop(event, n, sm = "PAS", method.ci = "WS"),
ma = FALSE, backtransf = FALSE)
print(metaprop(event, n, sm = "PRAW", method.ci = "WS"),
ma = FALSE, backtransf = FALSE)
# Same results (printed on original and log scale, respectively)
#
print(metaprop(event, n, sm = "PLN", method.ci = "NAsm"), ma = FALSE)
print(metaprop(event, n, sm = "PLN"), ma = FALSE, backtransf = FALSE)
# Results for first study (on log scale)
round(log(c(0.3079848, 0.2569522, 0.3691529)), 4)
# Print results as events per 1000 observations
#
print(metaprop(6:8, c(100, 1200, 1000), method = "Inverse"),
pscale = 1000, digits = 1)
# }
Run the code above in your browser using DataLab