H- and C-level WFST-based large vocabulary continuous speech recognition on Graphics Processing Units
School of Electrical Engineering, Seoul National University, San 45-1, Shillim-dong, Kwanak-gu, 151-744, Korea
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2011
@article{kim2011h,
title={H-AND C-LEVEL WFST-BASED LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION ON GRAPHICS PROCESSING UNITS},
author={Kim, J. and You, K. and Sung, W.},
booktitle={IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2011},
year={2011}
}
We have implemented 20,000-word large vocabulary continuous speech recognition (LVCSR) systems employing H- and C-level weighted finite state transducer (WFST) based networks on Graphics Processing Units (GPUs). Both the emission probability computation and the Viterbi beam search are implemented on the GPU in a data-parallel manner to minimize the extra data transfer time between the host CPU and the GPU. This study utilizes word-length optimization techniques to reduce the synchronization overhead in the Viterbi beam search. We achieve 18.6% to 21.9% of speed-up by using an efficient data packing method with less than 0.2% accuracy degradation. Furthermore, we explore different levels of abstraction in recognition network generation to reduce the number of synchronization operations as well as to minimize the memory usage. The experimental results show that the implemented systems on the GPU perform speech recognition 4.07 to 4.55 times faster than highly optimized sequential implementations on a CPU.
July 17, 2011 by hgpu