8258

Posts

Sep, 10

Accelerating Boosting-based Face Detection on GPUs

The goal of face detection is to determine the presence of faces in arbitrary images, along with their locations and dimensions. As it happens with any graphics workloads, these algorithms benefit from data-level parallelism. Existing parallelization efforts strictly focus on mapping different divide and conquer strategies into multicore CPUs and GPUs. However, even the most […]
Sep, 10

Performance Improvement of TOUGH2 Simulation with Graphics Processing Unit

We tried to accelerate the computational speed of TOUGH2 simulation by introducing a linear computation routine using a Graphics Processing Unit (GPU). Libraries for GPU computation were introduced, and new solvers for linear equations were developed. Out of those, CLLUSTB, an ILU preconditioned BiCGSTAB solver made with the CULA Sparse, demonstrated good performance both in […]
Sep, 8

Performance Evaluation of Concurrent Lock-free Data Structures on GPUs

Graphics processing units (GPUs) have emerged as a strong candidate for high-performance computing. While regular data-parallel computations with little or no synchronization are easy to map on the GPU architectures, it is a challenge to scale up computations on dynamically changing pointer-linked data structures. The traditional lock-based implementations are known to offer poor scalability due […]
Sep, 8

Lost in Translation: Challenges in Automating CUDA-to-OpenCL Translation

The use of accelerators in high-performance computing is increasing. The most commonly used accelerator is the graphics processing unit (GPU) because of its low cost and massively parallel performance. The two most common programming environments for GPU accelerators are CUDA and OpenCL. While CUDA runs natively only on NVIDIA GPUs, OpenCL is an open standard […]
Sep, 8

Mastering Software Variant Explosion for GPU Accelerators

Mapping algorithms in an efficient way to the target hardware poses a challenge for algorithm designers. This is particular true for heterogeneous systems hosting accelerators like graphics cards. While algorithm developers have profound knowledge of the application domain, they often lack detailed insight into the underlying hardware of accelerators in order to exploit the provided […]
Sep, 8

Supporting Heterogenous Computing Environments in SaC

From laptops to supercomputer nodes hardware architectures become increasingly heterogeneous, combining at least multiple general-purpose cores with one or even multiple GPGPU accelerators. Taking effective advantage of such systems’ capabilities becomes increasingly challenging. SaC is a functional array programming language with support for fully automatic parallelization following a data-parallel approach. As such many SaC programs […]
Sep, 8

OpenACC Implementations Comparison

Using GPUs for general purpose programming is, nowadays, much easier than the previous years. In the very beginning were Brook-GPU or Close To Metal the approaches used for exploring the new possibilities of hardware accelerators. After that, CUDA and OpenCL were released. They had been adopted by many programmers due to theirs advantages but, however, […]
Sep, 7

Can GPGPU Programming Be Liberated from the Data-Parallel Bottleneck?

Heterogeneous parallel primitives (HPP) addresses two major shortcomings in current GPGPU programming models: it supports full composability by defining abstractions and increases flexibility in execution by introducing braided parallelism.
Sep, 7

High Performance Parallel Implementation of Compressive Sensing SAR Imaging

The compressive sensing (CS) theory has been applied to SAR imaging systems in many ways. And it shows a significant reduction in the amount of sampling data at the cost of much longer reconstruction time. In this paper, we investigate the development and optimization of Iterative Shrinkage/Thresholding (IST) algorithm applying to CS reconstruction of SAR […]
Sep, 7

Multi-Moment Methods for PDEs and GPUs for Large-Scale Scientific Computations

The scope of this thesis is broad but focuses on developing effective numerical methods and efficient implementations. We investigate numerical solution methods for hyperbolic partial differential equations, numerical optimization methods, and implementation of fast numerical algorithms on graphics processor units (GPUs). For partial differential equations we develop numerical methods for the transport and advection equations. […]
Sep, 7

An Efficient Acceleration of Digital Fonensics Search Using GPGPU

Graphics Processing Units (GPU) have been the extensive research topic in recent years and have been successfully applied to general purpose applications other than computer graphical area. The nVidia CUDA programming model provides a straightforward means of describing inherently parallel computations. In this paper, we present a study of the efficiency of emerging technology in […]
Sep, 7

GPU Computing and CUDA technology used to accelerate a mesh generator application

The potential of GPU computing used in general purpose parallel programming has been amply shown. These massively parallel many-core multiprocessors are available to any users in every PCs, notebook, game console or workstation. In this work, we present the parallel version of a mesh-generating algorithm and its execution time reduction by using off-the-shelf GPU technology. […]

* * *

* * *

HGPU group © 2010-2025 hgpu.org

All rights belong to the respective authors

Contact us: