The number of space objects such as satellites, spacecraft, and debris are increasing significantly, and so is the need for tracking them for security and collision avoidance purposes. In this context, as parallelism is becoming a new paradigm, the need of implementing high performance propagators remain unmet. For this, we implemented Simplified General Perturbations No. […]

March 10, 2014 by hgpu

The aim of my thesis is to parallelize the Weighting Histogram Analysis Method (WHAM), which is a popular algorithm used to calculate the Free Energy of a molecular system in Molecular Dynamics simulations. WHAM works in post processing in cooperation with another algorithm called Umbrella Sampling. Umbrella Sampling has the purpose to add a biasing […]

February 2, 2014 by hgpu

The use of linear algebra routines is fundamental to many areas of computational science, yet their implementation in software still forms the main computational bottleneck in many widely used algorithms. In machine learning and computational statistics, for example, the use of Gaussian distributions is ubiquitous, and routines for calculating the Cholesky decomposition, matrix inverse and […]

January 29, 2014 by hgpu

Real time processing for teamwork action recognition is a challenge, due to complex computational models to achieve high system performance. Hence, this paper proposes a framework based on Graphical Processing Units (GPUs) to achieve a significant speed up in the performance of role based activity recognition of teamwork. The framework can be applied in various […]

October 15, 2013 by hgpu

It is important to obtain the results of methods that are used in solving scientific and engineering problems rapidly for users and application developers. Parallel programming techniques have been developed alongside serial programming because the importance of performance has been increasing day by day while developing computer applications.Various methods such as Gauss Elimination (GE) Method, […]

January 7, 2013 by hgpu

Parallel computing can offer an enormous advantage regarding the performance for very large applications in almost any field: scientific computing, computer vision, databases, data mining, and economics. GPUs are high performance many-core processors that can obtain very high FLOP rates. Since the first idea of using GPU for general purpose computing, things have evolved and […]

January 7, 2013 by hgpu

In the paper the result of application of artificial neural networks (ANN) for radionuclides transport modelling with surface runoff is presented. ANN with supervised training based on back propagation algorithm was used to predict radionuclides transport in the soil and on its surface. Application of ANN for substances migration modelling is worth using, because it […]

December 29, 2012 by hgpu

The change in processor architectures and 3D benchmarks makes performance characterization important for every processor and 3D application generation. Recent 3D applications require large amount of data to be processed by the GPU and the CPU. This leads to the importance in analyzing processor performance for different architectures and benchmarks so that benchmarks and processors […]

November 25, 2012 by hgpu

Multidimensional scaling (MDS) is a very popular and reliable method used in feature extraction and visualization of multidimensional data. The role of MDS is to reconstruct the topology of an original N-dimensional feature space consisting of M feature vectors in target 2-D (3-D) Euclidean space. It can be achieved by minimization of the error – […]

October 6, 2012 by hgpu

We present a parallel implementation for Graphics Processing Units (GPUs) of a model based on cellular automata (CA) to simulate laser dynamics. A cellular automaton is an inherent parallel type of algorithm that is very suitable to simulate complex systems formed by many individual components which give rise to emergent behaviours. We exploit the parallel […]

July 23, 2012 by hgpu

The exponential growth in user and application data entails new means for providing fault tolerance and protection against data loss. High Performance Computing (HPC) storage systems, which are at the forefront of handling the data deluge, typically employ hardware RAID at the backend. However, such solutions are costly, do not ensure end-to-end data integrity, and […]

July 3, 2012 by hgpu

In this paper, a new hybrid general-purpose computation on GPU (GPGPU) and computer graphics synthetic aperture radar (SAR) simulation method for complex scenes is proposed. Previous SAR simulations for complex scenes only use GPU’s graphics capabilities for scattering calculation in graphical electromagnetic computing (GRECO) algorithm. The new hybrid method use GPU’s graphics and parallel computing […]

March 12, 2012 by hgpu