Open Journal of Statistics
Vol.04 No.11(2014), Article ID:52713,9 pages
10.4236/ojs.2014.411083
Probit Normal Correlated Topic Model
Xingchen Yu, Ernest Fokoué
Center for Quality and Applied Statistics, Rochester Institute of Technology, Rochester, NY, USA
Email: xvy5021@gmail.com, ernest.fokoue@rit.edu
Copyright © 2014 by authors and Scientific Research Publishing Inc.
This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/



Received 3 October 2014; revised 28 October 2014; accepted 15 November 2014
ABSTRACT
The logistic normal distribution has recently been adapted via the transformation of multivariate Gaussian variables to model the topical distribution of documents in the presence of correlations among topics. In this paper, we propose a probit normal alternative approach to modelling correlated topical structures. Our use of the probit model in the context of topic discovery is novel, as many authors have so far concentrated solely of the logistic model partly due to the formidable inefficiency of the multinomial probit model even in the case of very small topical spaces. We herein circumvent the inefficiency of multinomial probit estimation by using an adaptation of the diagonal orthant multinomial probit in the topic models context, resulting in the ability of our topic modeling scheme to handle corpuses with a large number of latent topics. An additional and very important benefit of our method lies in the fact that unlike with the logistic normal model whose non-conjugacy leads to the need for sophisticated sampling schemes, our approach exploits the natural conjugacy inherent in the auxiliary formulation of the probit model to achieve greater simplicity. The application of our proposed scheme to a well-known Associated Press corpus not only helps discover a large number of meaningful topics but also reveals the capturing of compellingly intuitive correlations among certain topics. Besides, our proposed approach lends itself to even further scalability thanks to various existing high performance algorithms and architectures capable of handling millions of documents.
Keywords:
Topic Model, Bayesian, Gibbs Sampler, Cumulative Distribution Function, Probit, Logit, Diagonal Orthant, Efficient Sampling, Auxiliary Variable, Correlation Structure, Topic, Vocabulary, Conjugate, Dirichlet, Gaussian

1. Introduction
The task of recovering the latent topics underlying a given corpus of documents has been in the forefront of active research in statistical machine learning for more than a decade, and continues to receive the dedicated contributions from many researchers from around the world. Since the introduction of Latent Dirichlet Allocation (LDA) [1] and then the extension to correlated topic models (CTM) [2] , a series of excellent contributions have been made to this exciting field, ranging from slight extension in the modelling structure to the development of scalable topic modeling algorithms capable of handling extremely large collections of documents, as well as selecting an optimal model among a collection of competing models or using the output of topic modeling as entry points (inputs) to other machine learning or data mining tasks such as image analysis and sentiment extraction, just to name a few. As far as correlated topic models are concerned, virtually all the contributors to the field have so far concentrated solely on the use of the logistic normal topic model. The seminal paper on correlated topic model [2] adopts a variational approximation approach to model fitting while subsequent authors like [3] propose a Gibbs sampling scheme with data augmentation of uniform random variables. More recently, [4] presented an exact and scalable Gibbs sampling algorithm with Polya-Gamma distributed auxiliary variables which is a recent development of efficient sampling of logistic model. Despite the inseparable relationship between logistic and probit model in statistical modelling, the probit model has not yet been proposed, probably due to its computational inefficiency for multiclass classification problem and high posterior dependence between auxiliary variables and parameters. As for practical application where topic models are commonly employed, having multiple topics is extremely prevalent. In some cases, more than 1000 topics will be fitted to large datasets such as Wikipedia and Pubmed data. Therefore, using MCMC probit model in topic modeling application will be impractical and inconceivable due to its computational inefficiency. Nonetheless, a recent work on diagonal orthant probit model [5] substantially improved the sampling efficiency while maintaining the predictive performance, which motivated us to build an alternative correlated topic modeling with probit normal topic distribution. On the other hand, probit models inherently capture a better dependency structure between topics and co-occurrence of words within a topic as it doesn’t assume the IIA (independence of irrelevant alternatives) restriction of logistic models.
The rest of this paper is organized as follows: in Section 2, we present a conventional formulation of topic modeling along with our general notation and the correlated topic models extension. Section 3 introduces our adaptation of the diagonal orthant probit model to topic discovery in the presence correlations among topics, along with the corresponding auxiliary variable sampling scheme for updating the probit model parameters and the remainder of all the posterior distributions of the parameters of the model. Unlike with the logistic normal formulation where the non-conjugacy leads to the need for sophisticated sampling scheme, in this section we clearly reveal the simplicity of our proposed method resulting from the natural conjugacy inherent in the auxiliary formulation of the updating of the parameters. We also show compelling computational demonstrations of the efficiency of the diagonal orthant approach compared to the traditional multinomial probit for on both the auxiliary variable sampling and the estimation of the topic distribution. Section 4 presents the performance of our proposed approach on the Associated Press data set, featuring the intuitively appealing topics discovered, along with the correlation structure among topics and the loglikelihood as a function of topical space dimension. Section 5 deals with our conclusion, discussion and elements of our future work.
2. General Aspects of Topic Models
In a given corpus, one could imagine that each document deals with one or more topics. For instance, one of the collection considered in this paper is provided by the Associated Press and covers topics as varied as aviation, education, weather, broadcasting, air force, navy, national security, international treaties, investing, interna- tional trade, war, courts, entertainment industry, politics, and etc. From a statistical perspective, a topic is often modeled as a probability distribution over words, and as a result a given document is treated as a mixture of probabilistic topics [1] . We consider a setting where we have a total of
unique words in the reference vocabulary and
topics underlying the
documents provided. Let
denote the n-th word in the d-th document, and let
refer to the label of the topic assigned to the
-th word of that
-th document. Then the probability of
is given by
(1)
where
is the probability that the n-th word in the d-th document is assigned to topic
. This quantity plays an important role in the analysis of correlated topic models. In the seminal article on correlated topic models [2] ,
is modeled for each document
as a function of a K-dimensional vector 
of parameters. Specifically, the logistic-normal defines
where the last element
is typically set to zero for identifiability and assumes with 
Also, 



where 



(3) provides the ingredients for estimating the parameter vectors 



The practical evaluation of (3) involves a complicated high dimensional integral which is typically computationally intractable when the number of categories is greater than 4. A relaxed version of (3), one that still captures more correlation than the logit and that is also very commonly used in practice, defines 

where 

tribution function. Despite this relaxation, the multinomial probit in this formulation still has major drawbacks namely: 1) Even when one is given the vector















The condition in (5) typically fails to be fulfilled even when 



3. Diagonal Orthant Probit for Correlated Topic Models
In a recent work, [5] developed the diagonal orthant probit approach to multicategorical classification. Their approach circumvents the bottlenecks mentioned earlier and substantially improves the sampling efficiency while maintaining the predictive performance. Essentially, the diagonal orthant probit approach successfully makes the most of the benefits of binary classification, thereby substantially reducing the high dependency that made the condition (5) computationally unattainable. Indeed, with the diagonal orthant multinomial model, we achieved three main benefits
・ A more tractable and easily computatble definition of topic distribution
・ A clear and very straightforward and adaptable auxiliary variable sampling scheme
・ Thecapacity to handle a very large number of topics due to the efficiency and low dependency.
Under the diagonal orthant probit model, we have

The generative process of our probit normal topic models is essentially identical to logistic topic models except that the topic distribution for each document now is obtained by a probit transformation of a multivariate Gaussian variable (6). As such, the generating process of a document of length 
1) Draw 




2) For each word position
a) Draw a topic assignment
b) Draw a word
Where 




To complete the Bayesian analysis of our probit normal topic model, we need to sample from the joint posterior

As noted earlier, the second benefit of the diagonal orthant probit model lies in its clear, simple, straightforward yet powerful auxiliary variable sampling scheme. We take advantage of that diagonal orthant property when dealing with the full posterior for 

While sampling directly from (9) is impractical, defining a collection of auxiliary variables 

For each document

Each row 


readily using the following straightforward sampling scheme: Let 
・ For the component of 

・ For all components of 

Once the matrix 

where
with 




meaning that 

where


where
As far as sampling from the full posterior distribution of 
where the use of 
4. Computational Results on the Associated Press Data
In this section, we used a famous Associated Press data set from [7] in R to uncover the word topic distribution, the correlation structure between various topics as well as selecting optimal models. The Associated Press corpus consists of 2244 documents and 10,473 words. After preprocessing the corpus by picking frequent and common terms, we reduced the size of the words from 10,473 to 2643 for efficient sampling.
In our first experimentation, we built a correlated topic modeling structure based on the traditional multinomial probit and then tested the computational speed for key sampling tasks. The high posterior dependency structure between auxiliary variables and parameters make multinormal probit essentially unscalable for situations where it is impossible for the sampler to yield a random variate of the auxiliary variable corresponding the current topic allocation label that is also the maximum (5). For a random initialization of topic assignment, the sampling of auxiliary variable cannot even complete one single iteration. In the case of good initialization of topical prior 



Table 1. All the numbers in this table represent the processing time (in se- conds), and are computed in Ron PC using a parallel algorithm acting on 4 CPU cores. NA here represents situations where it is impossible for the sampler to yield a random variate of the auxiliary variable corresponding the current topic allocation label that is also the maximum.
In addition to the drastic improvement of the overall sampling efficiency, we noticed that the computational complexity for sampling the auxiliary variable and topic distribution is close to O(1) and O(K) respectively, suggesting that probit normal topic model now becomes an attainable and feasible tool of the traditional correlated topic model.
Central to topic modeling is the need to determine for a given corpus the optimal number of latent topics. As it is the case for most latent variable models, this task can be formidable at times, and there is no consensus among machine learning researchers as to which of the existing methods is the best. Figure 1 shows the loglikelihood as a function of the number of topics discovered in the model. Apart from the loglikelihood, many other techniques are commonly used such as perplexity, harmonic mean method and so on.
As we see, the optimal number of topics in this case is 30. In Table 2, we show a subset of the 30 topics uncovered where each topic is represented by the 10 most frequent words. It can be seen that our probit normal topic model is able to capture the co-occurrence of words within topics successfully. In Figure 2, we also show the correlation structure between various topics which is the essential purpose of employing the correlated topic model. Evidently, the correlation captured intuitively reflect the natural relationship between similar topics.
5. Conclusion and Discussion
In the context of topic modeling where many other researchers seem to have avoided it. By adapting the diagonal orthant probit model, we proposed a probit alternative to the logit approach to the topic modeling. Compared to the multinomial probit model we constructed, our topic discovery scheme using diagonal orthant probit model enjoyed several desirable properties; First,we gained the efficiency in computing the topic distribution
In the Associated Press example explored in the previous section, not only does our method produce a better likelihood than the logistic normal topic model with variational EM, but also discovers meaningful topics along with underlying correlation structure between topics. Overall, the method we developed in this paper offers another feasible alternatives in the context of correlated topic model that we hope will be further explored and extended by many other researchers.
Figure 1. Log likelihood as a function of the number of topics.
Table 2. Representation of topics discovered by our method.
Based on the promising results we have seen in this paper, the probit normal topic model opens the door for various future works. For instance, [8] proposed a multi-field correlated topic model by relaxing the assumption of using common set of topics globally among all documents, which can also be applied to the probit model to enrich the comprehensiveness of structural relationships between topics. Another potential direction would be to
Figure 2. Graphical representation of the correlation among topics.
enhance the scalability of the model. Currently we used a simple distributed algorithm proposed by [9] and [10] for efficient Gibbs sampling. The architecture for topic models presented by [11] can be further utilized to reduce the computational complexity substantially while delivering comparable performance. Furthermore, a novel sampling method involving the Gibbs Max-Margin Topic [12] will further improve the computational efficiency.
Acknowledgements
We want to express our sincere gratitude to our reviewer for comprehensive and constructive advice. We also wish to express our heartfelt gratitude and infinite thanks to Our Lady of Perpetual Help for Her ever-present support and guidance, especially for the uninterrupted flow of inspiration received through Her most powerful intercession.
References
- Blei, D.M. and Ng, A.Y., Jordan, M.I. and Lafferty, J. (2003) Latent Dirichlet Allocation. Journal of Machine Learning Research, 3.
- Blei, D.M. and Lafferty, J.D. (2006) Correlated Topic Models. Proceedings of the 23rd International Conference on Machine Learning, MIT Press, Cambridge, Massachusetts, 113-120.
- Mimno, D., Wallach, H.M. and Mccallum, A. (2008) Gibbs Sampling for Logistic Normal Topic Models with Graph- Based Priors. Proceedings of NIPS Workshop on Analyzing Graphs, 2008.
- Chen, J.F., Zhu, J., Wang, Z., Zheng, X. and Zhang, B. (2013) Scalable Inference for Logistic-Normal Topic Models. In Burges, C.J.C., Bottou, L., Welling, M., Ghahramani, Z. and Weinberger, K.Q., Eds., Advances in Neural Information Processing Systems 26, Curran Associates, Inc., 2445-2453.
- Johndrow, J., Lum, K. and Dunson, D.B. (2013) Diagonal Orthant Multinomial Probit Models. JMLR Proceedings, Volume 31 of AISTATS, 29-38.
- Albert, J.H. and Chib, S. (1993) Bayesian Analysis of Binary and Polychotomous Response Data. Journal of the American Statistical Association, 88, 669-679. http://dx.doi.org/10.1080/01621459.1993.10476321
- Grun, B. and Hornik, K. (2011) Topicmodels: An R Package for Fitting Topic Models. Journal of Statistical Software, 40, 1-30.
- Salomatin, K., Yang, Y.M. and Lad, A. (2009) Multi-Field Correlated Topic Modeling. Proceedings of the SIAM International Conference on Data Mining, SDM 2009, April 30-May 2, 2009, Sparks, 628-637.
- Yao, L.M., Mimno, D. and McCallum, A. (2009) Efficient Methods for Topic Model Inference on Streaming Document Collections. KDD 2009: Proceedings of 15th ACM SIGKDD int’l Conference on Knowledge Discovery and Data Mining, 937-946.
- Newman, D., Asuncion, A., Smyth, P. and Welling, M. (2009) Distributed Algorithms for Topic Models. Journal of Machine Learning Research, 10, 1801-1828.
- Smola, A. and Narayanamurthy, S. (2010) An Architecture for Parallel Topic Models. Proc. VLDB Endow., 3, 703-710.
- Zhu, J., Chen, N., Perkins, H. and Zhang, B. (2013) Gibbs Max-Margin Topic Models with Data Augmentation. CoRR, abs/1310.2816.





















