Journal
PROCEEDINGS OF THE VLDB ENDOWMENT
Volume 3, Issue 1, Pages 703-710Publisher
ASSOC COMPUTING MACHINERY
DOI: 10.14778/1920841.1920931
Keywords
-
Funding
- Australian Research Council
Ask authors/readers for more resources
This paper describes a high performance sampling architecture for inference of latent topic models on a cluster of workstations. Our system is faster than previous work by over an order of magnitude and it is capable of dealing with hundreds of millions of documents and thousands of topics. The algorithm relies on a novel communication structure, namely the use of a distributed (key, value) storage for synchronizing the sampler state between computers. Our architecture entirely obviates the need for separate computation and synchronization phases. Instead, disk, CPU, and network are used simultaneously to achieve high performance. We show that this architecture is entirely general and that it can be extended easily to more sophisticated latent variable models such as n-grams and hierarchies.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available