Parallel implementations of probabilistic latent semantic analysis on graphic processing units
Graduate College, University of Illinois, Urbana-Champaign
University of Illinois, 2011
@article{zhai2011parallel,
title={Parallel implementations of probabilistic latent semantic analysis on graphic processing units},
author={Zhai, C.X.},
year={2011}
}
Probabilistic Latent Semantic Analysis (PLSA) has been successfully applied to many text mining tasks such as retrieval, clustering, summarization, etc. PLSA involves iterative computation for a large number of parameters and may take hours or even days to process a large dataset, thus speeding up PLSA is highly motivated in the domain of text mining. Recently, the general purpose graphic processing units (GPGPU) have become a powerful parallel computing platform, not only because of GPU’s multi-core structure and high memory bandwidth, but also because of the recent efforts devoted into building a programming framework to enable developers to easily manipulate GPU’s computing power. In this paper, we introduced two methods to parallelize and speed up PLSA via GPGPU. Related issues are addressed including workload balance, block-thread layout, memory and data access optimization, etc. The GPU in use is NVidia GTX480 (costs $450 in market). Experimental results show that our methods can process 300,000 documents in 12 seconds which is a 33x speedup compared with traditional PLSA implementation running on 3.0GHz Intel Xeon CPU. The significant speedup can bring researchers in the text mining domain brand new experience.
September 27, 2011 by hgpu