14014

Posts

May, 19

An Interrupt-Driven Work-Sharing For-Loop Scheduler

In this paper we present a parallel for-loop scheduler which is based on work-stealing principles but runs under a completely cooperative scheme. POSIX signals are used by idle threads to interrupt left-behind workers, which in turn decide what portion of their workload can be given to the requester. We call this scheme Interrupt-Driven Work-Sharing (IDWS). […]
May, 18

A Survey Of Architectural Approaches for Data Compression in Cache and Main Memory Systems

As the number of cores on a chip increase and key applications become even more data-intensive, memory systems in modern processors have to deal with increasingly large amount of data. In face of such challenges, data compression presents as a promising approach to increase effective memory system capacity and also provide performance and energy advantages. […]
May, 18

Workshop on Heterogeneous and Unconventional Cluster Architectures and Applications (HUCAA2015), 2015

====================================================================== CALL FOR PAPERS 4th International Workshop on Heterogeneous and Unconventional Cluster Architectures and Applications (HUCAA 2015) http://www.hucaa-workshop.org/hucaa2015 Sept. 8-11, 2015 – Chicago, IL, US In conjunction with IEEE CLUSTER 2015 IEEE International Conference on Cluster Computing ====================================================================== ABOUT THE WORKSHOP The workshop on Heterogeneous and Unconventional Cluster Architectures and Applications gears to gather recent […]
May, 16

A Fast and Rigorously Parallel Surface Voxelization Technique for GPU-Accelerated CFD Simulations

This paper presents a fast surface voxelization technique for the mapping of tessellated triangular surface meshes to uniform and structured grids that provide a basis for CFD simulations with the lattice Boltzmann method (LBM). The core algorithm is optimized for massively parallel execution on graphics processing units (GPUs) and is based on a unique dissection […]
May, 16

Multi-GPU Support on Single Node Using Directive-Based Programming Model

Existing studies show that using single GPU can lead to obtaining significant performance gains. We should be able to achieve further performance speedup if we use more than one GPU. Heterogeneous processors consisting of multiple CPUs and GPUs offer immense potential and are often considered as a leading candidate for porting complex scientific applications. Unfortunately […]
May, 16

Efficient Resource Scheduling for Big Data Processing on Accelerator-based Heterogeneous Systems

The involvement of accelerators is becoming widespread in the field of heterogeneous processing, performing computation tasks through a wide range of applications. In this paper, we examine the heterogeneity in modern computing systems, particularly, how to achieve a good level of resource utilization and fairness, when multiple tasks with different load and computation ratios are […]
May, 16

Performance Analysis and Efficient Execution on Systems with multi-core CPUs, GPUs and MICs

We carry out a comparative performance study of multi-core CPUs, GPUs and Intel Xeon Phi (Many Integrated Core – MIC) with a microscopy image analysis application. We experimentally evaluate the performance of computing devices on core operations of the application. We correlate the observed performance with the characteristics of computing devices and data access patterns, […]
May, 16

Using Butterfly-Patterned Partial Sums to Optimize GPU Memory Accesses for Drawing from Discrete Distributions

We describe a technique for drawing values from discrete distributions, such as sampling from the random variables of a mixture model, that avoids computing a complete table of partial sums of the relative probabilities. A table of alternate ("butterfly-patterned") form is faster to compute, making better use of coalesced memory accesses. From this table, complete […]
May, 15

Speeding up Automatic Hyperparameter Optimization of Deep Neural Networks by Extrapolation of Learning Curves

Deep neural networks (DNNs) show very strong performance on many machine learning problems, but they are very sensitive to the setting of their hyperparameters. Automated hyperparameter optimization methods have recently been shown to yield settings competitive with those found by human experts, but their widespread adoption is hampered by the fact that they require more […]
May, 15

MRCUDA: MapReduce Acceleration Framework Based on GPU

GPU programming model for general purpose computing is complex and difficult to be maintained. A MapReduce acceleration framework named MRCUDA is designed and implemented in this paper. There are four loosely coupled stages in MRCUDA, including Pre-Processing, Map, Group and Reduce, which can support flexible configurations for different applications. In order to take full advantage […]
May, 15

The 3D Flow Field Around an Embedded Planet

Understanding the 3D flow topology around a planet embedded in its natal disk is crucial to the study of planet formation. 3D modifications to the well-studied 2D flow topology have the potential to resolve longstanding problems in both planet migration and accretion. We present a detailed analysis of the 3D isothermal flow field around a […]
May, 15

Adaptive discrete cosine transform-based image compression method on a heterogeneous system platform using Open Computing Language

Discrete cosine transform (DCT) is one of the major operations in image compression standards and it requires intensive and complex computations. Recent computer systems and handheld devices are equipped with high computing capability devices such as a general-purpose graphics processing unit (GPGPU) in addition to the traditional multicores CPU. We develop an optimized parallel implementation […]

* * *

* * *

HGPU group © 2010-2025 hgpu.org

All rights belong to the respective authors

Contact us: