Functions to Fit LDA-type models
These functions use a collapsed Gibbs sampler to fit three different models: latent Dirichlet allocation (LDA), the mixed-membership stochastic blockmodel (MMSB), and supervised LDA (sLDA). These functions take sparsely represented input documents, perform inference, and return point estimates of the latent parameters using the state at the last iteration of Gibbs sampling. Multinomial logit for sLDA is supported using the multinom function from nnet package .
lda.collapsed.gibbs.sampler(documents, K, vocab, num.iterations, alpha, eta, initial = NULL, burnin = NULL, compute.log.likelihood = FALSE, trace = 0L, freeze.topics = FALSE)slda.em(documents, K, vocab, num.e.iterations, num.m.iterations, alpha, eta, annotations, params, variance, logistic = FALSE, lambda = 10, regularise = FALSE, method = "sLDA", trace = 0L, MaxNWts=3000)mmsb.collapsed.gibbs.sampler(network, K, num.iterations, alpha, beta.prior, initial = NULL, burnin = NULL, trace = 0L)lda.cvb0(documents, K, vocab, num.iterations, alpha, eta, trace = 0L)
A list whose length is equal to the number of documents, D. Each
element of documents is an integer matrix with two rows. Each
column of documents[[i]] (i.e., document $i$) represents a
word occurring in the document.
documents[[i]][1, j] is a 0-indexed word identifier for the jth word in document i. That is, this should be an index - 1 into vocab. documents[[i]][2, j] is an integer specifying the number of times that word appears in the document.
mmsb.collapsed.gibbs.sampler, a $D \times D$ matrix (coercible as logical) representing the adjacency matrix for the network. Note that elements on the diagonal are ignored.
- An integer representing the number of topics in the model.
- A character vector specifying the vocabulary words associated with the word indices used in documents.
- The number of sweeps of Gibbs sampling over the entire corpus to make.
slda.em, the number of Gibbs sampling sweeps to make over the entire corpus for each iteration of EM.
slda.em, the number of EM iterations to make.
- The scalar value of the Dirichlet hyperparameter for topic proportions.
mmsb.collapsed.gibbs.sampler, the the beta hyperparameter for each entry of the block relations matrix. This parameter should be a length-2 list whose entries are $K \times K$ matrices. The elements of the two matrices comprise the two parameters for each beta variable.
- The scalar value of the Dirichlet hyperparamater for topic multinomials.
- A list of initial topic assignments for words. It should be in the same format as the assignments field of the return value. If this field is NULL, then the sampler will be initialized with random assignments.
A scalar integer indicating the number of Gibbs sweeps to consider
as burn-in (i.e., throw away) for
mmsb.collapsed.gibbs.sampler. If this parameter is non-NULL, it will also have the side-effect of enabling the document_expects field of the return value (see below for details). Note that burnin iterations do NOT count towards num.iterations.
A scalar logical which when
TRUEwill cause the sampler to compute the log likelihood of the words (to within a constant factor) after each sweep over the variables. The log likelihood for each iteration is stored in the log.likelihood field of the result. This is useful for assessing convergence, but slows things down a tiny bit.
A length D numeric vector of covariates associated with each
document. Only used by
slda.emwhich models documents along with numeric annotations associated with each document. When using the logistic option, annotations must be consecutive integers starting from 0.
slda.em, a length Kx(number of classes-1) numeric vector of regression coefficients at which the EM algorithm should be initialized.
slda.em, the variance associated with the Gaussian response modeling the annotations in annotations.
slda.em, a scalar logical which, when
TRUE, causes the annotations to be modeled using a logistic response instead of a Gaussian (the covariates must be consecutive integers starting from zero when used with sLDA).
When regularise is
TRUE. This is a scalar that is the standard deviation of the Gaussian prior on the regression coefficients.
TRUE, a Gaussian prior is used for the regression coefficients. This requires the
slda.em, a character indicating how to model the annotations. Only
"sLDA", the stock model given in the references, is officially supported at the moment.
traceis greater than zero, diagnostic messages will be output. Larger values of
traceimply more messages.
- Input to the nnet's multinom function with a default value of 3000 maximum weights. Increasing this value may be necessary when using logistic sLDA with a large number of topics at the necessary expense of longer run times.
TRUE, topic assignments will occur but the counts of words associated with topics will not change. initial should be set when this option is used. This is best use for sampling test documents.
A fitted model as a list with the following components:
- A list of length D. Each element of the list, say
assignments[[i]]is an integer vector of the same length as the number of columns in
documents[[i]]indicating the topic assignment for each word.
- A $K \times V$ matrix where each entry indicates the number of times a word (column) was assigned to a topic (row). The column names should correspond to the vocabulary words given in vocab.
- A length K vector where each entry indicates the total number of times words were assigned to each topic.
- A $K \times D$ matrix where each entry is an integer indicating the number of times words in each document (column) were assigned to each topic (column).
- Only for
lda.collapsed.gibbs.sampler. A matrix with 2 rows and
num.iterationscolumns of log likelihoods when the flag
compute.log.likelihoodis set to
TRUE. The first row contains the full log likelihood (including the prior), whereas the second row contains the log likelihood of the observations conditioned on the assignments.
- This field only exists if burnin is non-NULL. This field is like document_sums but instead of only aggregating counts for the last iteration, this field aggegates counts over all iterations after burnin.
- Only for
mmsb.collapsed.gibbs.sampler. A $D \times D$ integer matrix of topic assignments for the source document corresponding to the link between one document (row) and another (column).
- Only for
mmsb.collapsed.gibbs.sampler. A $D \times D$ integer matrix of topic assignments for the destination document corresponding to the link between one document (row) and another (column).
- Only for
mmsb.collapsed.gibbs.sampler. A $K \times K$ integer matrix indicating the number of times the source of a non-link was assigned to a topic (row) and the destination was assigned to another (column).
- Only for
mmsb.collapsed.gibbs.sampler. A $K \times K$ integer matrix indicating the number of times the source of a link was assigned to a topic (row) and the destination was assigned to another (column).
slda.em, a model of type
lm, the regression model fitted to the annotations.
slda.em, a length Kx(number of classes-1) numeric vector of coefficients for the regression model.
WARNING: This function does not compute precisely the correct thing when the count associated with a word in a document is not 1 (this is for speed reasons currently). A workaround when a word appears multiple times is to replicate the word across several columns of a document. This will likely be fixed in a future version.
Blei, David M. and Ng, Andrew and Jordan, Michael. Latent Dirichlet allocation. Journal of Machine Learning Research, 2003.
Airoldi , Edoardo M. and Blei, David M. and Fienberg, Stephen E. and Xing, Eric P. Mixed Membership Stochastic Blockmodels. Journal of Machine Learning Research, 2008.
Blei, David M. and McAuliffe, John. Supervised topic models. Advances in Neural Information Processing Systems, 2008.
Griffiths, Thomas L. and Steyvers, Mark. Finding scientific topics. Proceedings of the National Academy of Sciences, 2004.
Asuncion, A., Welling, M., Smyth, P., and Teh, Y. W. On smoothing and inference for topic models. Uncertainty in Artificial Intelligence, 2009.
## See demos for the three functions: ## Not run: demo(lda) ## Not run: demo(slda) ## Not run: demo(mmsb)