A Distributed CPU-GPU Framework for Pairwise Alignments on Large-Scale Sequence Datasets
Dept. of Electrical and Computer Engineering, University of Missouri
24th IEEE International Conference on Application-specific Systems, Architectures and Processors (ASAP), 2013
@article{da2013distributed,
title={A Distributed CPU-GPU Framework for Pairwise Alignments on Large-Scale Sequence Datasets},
author={Da Li, Kittisak Sajjapongse and Truong, Huan and Conant, Gavin and Becchi, Michela},
year={2013}
}
Several problems in computational biology require the all-against-all pairwise comparisons of tens of thousands of individual biological sequences. Each such comparison can be performed with the well-known Needleman-Wunsch alignment algorithm. However, with the rapid growth of biological databases, performing all possible comparisons with this algorithm in serial becomes extremely time-consuming. The massive computational power of graphics processing units (GPUs) makes them an appealing choice for accelerating these computations. As such, CPU-GPU clusters can enable all-against-all comparisons on large datasets. In this paper, we present a hybrid MPI-CUDA framework for computing multiple pairwise sequence alignments on CPU-GPU clusters. Our design targets both homogeneous and heterogeneous clusters with nodes characterized by different hardware and computing capabilities. Our framework consists of the following components: a cluster-level dispatcher, a set of node-level dispatchers, and a set of CPU- and GPU-workers. The cluster-level dispatcher progressively distributes work to the compute nodes and aggregates the results. The node-level dispatchers distribute alignment tasks to available CPUs and GPUs and perform dual-buffering to hide data transfers between CPU and GPU. CPU- and GPU-workers perform pairwise sequence alignments using the Needleman-Wunsch algorithm. We propose and evaluate three designs for these GPU workers, all of them outperforming the existing open-source implementation from the Rodinia Benchmark Suite.
May 11, 2013 by hgpu