Fast, Processor-Cardinality Agnostic PRNG with a Tracking Application

Andrew Janowczyk, Sharat Chandran, Srinivas Aluru
ViGIL, Department of Computer Science and Engineering, Indian Insitute of Technology Bombay, Mumbai
Sixth Indian Conference on Computer Vision, Graphics & Image Processing, 2008. ICVGIP ’08


   title={Fast, processor-cardinality agnostic PRNG with a tracking application},

   author={Janowczyk, A. and Chandran, S. and Aluru, S.},

   booktitle={Computer Vision, Graphics & Image Processing, 2008. ICVGIP’08. Sixth Indian Conference on},





Download Download (PDF)   View View   Source Source   



As vision algorithms mature with increasing inspiration from the learning community, statistically independent pseudo random number generation (PRNG) becomes increasingly important. At the same time, execution time demands have seen algorithms being implemented on evolving parallel hardware such as GPUs. The Mersenne Twister (MT) has proven to be the current state of the art for generating high quality random numbers, and the Nvidia provided software for parallel MT is in widespread use. While execution time is important, development time is also critical. As processor cardinality changes, a foundation for generating simulations that will vary only in execution time and not in the actual result is useful; otherwise the development time will be impacted. In this paper, we present an implementation of the Lagged Fibonacci Generator (LFG) considered to be of quality equal to MT on the GPU. Unlike MT, LFG has this important processor-cardinality agnostic capability — that is — as the number of processing resources changes, the overall sequence of random numbers remains the same. This feature not withstanding, our basic implementation is roughly as fast as the parallel MT; an in-memory version is actually 25% faster in execution time. Both parallel MT as well as parallel LFG show enormous speed up over their sequential counterparts. Finally, a prototype particle filter tracking application shows that our method works not just in parallel computing theory, but also in practice for vision applications, providing a decrease of 60% in execution time.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2020 hgpu.org

All rights belong to the respective authors

Contact us: