Blesson Varghese, Javier Prades, Carlos Reano, Federico Silla
‘How can GPU acceleration be obtained as a service in a cluster?’ This question has become increasingly significant due to the inefficiency of installing GPUs on all nodes of a cluster. The research reported in this paper is motivated to address the above question by employing rCUDA (remote CUDA), a framework that facilitates Acceleration-as-a-Service (AaaS), […]
View View   Download Download (PDF)   
Gordon Inggs, David B. Thomas, Wayne Luk
Users of heterogeneous computing systems face two problems: firstly, understanding the trade-off relationship between the observable characteristics of their applications, such as latency and quality of the result, and secondly, how to exploit knowledge of these characteristics to allocate work to distributed resources efficiently. A domain specific approach addresses both of these problems. By considering […]
View View   Download Download (PDF)   
Joris Cramwinckel
In this thesis we present a state-of-the-art approach to accelerate Monte Carlo valuations of embedded options. Due to regulations and improved risk management, nested simulations (scenarios in scenarios) are becoming increasingly important for institutional investors like: insurance companies, pension funds and housing corporations. Preferably one wishes to use a framework in which multiple related problems […]
View View   Download Download (PDF)   
Christian Andreetta, Vivien Begot, Jost Berthold, Martin Elsman, Troels Henriksen, Maj-Britt Nordfang, Cosmin E. Oancea
Commodity many-core hardware is now mainstream, driven in particular by the evolution of general purpose graphics programming units (GPGPUs), but parallel programming models are lagging behind in effectively exploiting the available application parallelism. There are two principal reasons. First, real-world applications often exhibit a rich composition of nested parallelism, whose statical extraction requires a set […]
View View   Download Download (PDF)   
Blesson Varghese
The risk of reinsurance portfolios covering globally occurring natural catastrophes, such as earthquakes and hurricanes, is quantified by employing simulations. These simulations are computationally intensive and require large amounts of data to be processed. The use of many-core hardware accelerators, such as the Intel Xeon Phi and the NVIDIA Graphics Processing Unit (GPU), are desirable […]
View View   Download Download (PDF)   
Giorgis Georgakoudis, Charles J. Gillan, Ahmed Sayed, Ivor Spence, Richard Faloon, Dimitrios S. Nikolopoulos
Energy efficiency has been a daunting challenge for datacenters. The financial industry operates some of the largest datacenters in the world. With increasing energy costs and the financial services sector growth, emerging financial analytics workloads may incur extremely high operational costs, to meet their latency targets. Microservers have recently emerged as an alternative to high-end […]
View View   Download Download (PDF)   
Di Zhao
High-accuracy optimizer is the essential part of accuracy-sensitive applications such as computational finance and computational biology, and we developed single-GPU based Iterative Discrete Approximation Monte Carlo Search (IDA-MCS) in our previous research. However, single-GPU IDA-MCS is in low performance or even functionless for optimization problems with large number of peaks because of the capability constrains […]
View View   Download Download (PDF)   
Moritz Kreutzer, Georg Hager, Gerhard Wellein, Andreas Pieper, Andreas Alvermann, Holger Fehske
The Kernel Polynomial Method (KPM) is a well-established scheme in quantum physics and quantum chemistry to determine the eigenvalue density and spectral properties of large sparse matrices. In this work we demonstrate the high optimization potential and feasibility of peta-scale heterogeneous CPU-GPU implementations of the KPM. At the node level we show that it is […]
View View   Download Download (PDF)   
Robin Kumar, Amandeep Kaur Cheema
Machine learning, a branch of artificial intelligence, concerns the construction and study of systems that can learn from data. Neural network is the well-known branch of machine learning & it has been used extensively by researchers for prediction of data and the prediction accuracy depends upon fine tuning of particular financial data. In this paper […]
View View   Download Download (PDF)   
Chuan-Hsiang Han, Yu-Tuan Lin
Monte Carlo simulations have become widely used in computational finance. Standard error (SE in short) is the basic notion to measure the quality of a Monte Carlo estimator, and the square of SE is defined as the variance divided by the total number of simulations. Variance reduction methods have been developed as efficient algorithms by […]
View View   Download Download (PDF)   
Linlin Xu, Giray Okten
GPU computing has become popular in computational finance and many financial institutions are moving their CPU based applications to the GPU platform. Since most Monte Carlo algorithms are embarrassingly parallel, they benefit greatly from parallel implementations, and consequently Monte Carlo has become a focal point in GPU computing. GPU speed-up examples reported in the literature […]
View View   Download Download (PDF)   
Gordon Inggs, David Thomas, Wayne Luk
We advocate a domain specific software development methodology for heterogeneous computing platforms such as Multicore CPUs, GPUs and FPGAs. We argue that three specific benefits are realised from adopting such an approach: portable, efficient implementations across heterogeneous platforms; domain specific metrics of quality that characterise platforms in a form software developers will understand; automatic, optimal […]
Page 1 of 1012345...10...Last »

* * *

* * *

Follow us on Twitter

HGPU group

1587 peoples are following HGPU @twitter

Like us on Facebook

HGPU group

303 people like HGPU on Facebook

* * *

Free GPU computing nodes at hgpu.org

Registered users can now run their OpenCL application at hgpu.org. We provide 1 minute of computer time per each run on two nodes with two AMD and one nVidia graphics processing units, correspondingly. There are no restrictions on the number of starts.

The platforms are

Node 1
  • GPU device 0: nVidia GeForce GTX 560 Ti 2GB, 822MHz
  • GPU device 1: AMD/ATI Radeon HD 6970 2GB, 880MHz
  • CPU: AMD Phenom II X6 @ 2.8GHz 1055T
  • RAM: 12GB
  • OS: OpenSUSE 13.1
  • SDK: nVidia CUDA Toolkit 6.5.14, AMD APP SDK 3.0
Node 2
  • GPU device 0: AMD/ATI Radeon HD 7970 3GB, 1000MHz
  • GPU device 1: AMD/ATI Radeon HD 5870 2GB, 850MHz
  • CPU: Intel Core i7-2600 @ 3.4GHz
  • RAM: 16GB
  • OS: OpenSUSE 12.3
  • SDK: AMD APP SDK 3.0

Completed OpenCL project should be uploaded via User dashboard (see instructions and example there), compilation and execution terminal output logs will be provided to the user.

The information send to hgpu.org will be treated according to our Privacy Policy

HGPU group © 2010-2015 hgpu.org

All rights belong to the respective authors

Contact us: