10865

Discrete Shearlet Transform on GPU with Applications in Anomaly Detection and Denoising

Xavier Gibert Serra, Vishal M. Patel, Demetrio Labate, Rama Chellappa
UMIACS, University of Maryland, College Park, MD 20742-3275, USA
University of Maryland, 2013
@article{serra2013discrete,

   title={Discrete Shearlet Transform on GPU with Applications in Anomaly Detection and Denoising},

   author={Serra, Xavier Gibert and Patel, Vishal M. and Labate, Demetrio and Chellappa, Rama},

   year={2013}

}

Download Download (PDF)   View View   Source Source   Source codes Source codes

Package:

730

views

Shearlets have emerged in recent years as one of of the most successful methods for the multiscale analysis of multidimensional signals. Unlike wavelets, shearlets form a pyramid of well-localized functions defined not only over a range of scales and locations, but also over a range of orientations and with highly anisotropic supports. As a result, shearlets are much more effective than traditional wavelets in handling the geometry of multidimensional data and this was exploited in a wide range of applications from image and signal processing. However, despite their desirable properties, the wider applicability of shearlets is limited by its computational complexity. For example, denoising a single 512 512 image using a current implementation of the shearlet-based shrinkage algorithm can take between 10 seconds and 2 minutes, depending on the number of CPU cores, and much longer processing times are required for video denoising. However, due to the parallel nature of the shearlet transform, it is possible to use Graphical Processing Units (GPU) to accelerate it. We provide an open source standalone implementation of the 2D shearlet using CUDA C++ as well as GPU-accelerated MATLAB implementations of the 2D and 3D shearlet transforms. We have instrumented the code so that we can analyze the running time of each kernel under different GPU hardware. In addition to denoising, we describe a novel application of shearlets for detecting anomalies on textured images. In this application, computation times can be reduced by a factor of 50 or more, compared to multi-core CPU implementations.
VN:F [1.9.22_1171]
Rating: 0.0/5 (0 votes cast)

* * *

* * *

Like us on Facebook

HGPU group

184 people like HGPU on Facebook

Follow us on Twitter

HGPU group

1311 peoples are following HGPU @twitter

* * *

Free GPU computing nodes at hgpu.org

Registered users can now run their OpenCL application at hgpu.org. We provide 1 minute of computer time per each run on two nodes with two AMD and one nVidia graphics processing units, correspondingly. There are no restrictions on the number of starts.

The platforms are

Node 1
  • GPU device 0: AMD/ATI Radeon HD 5870 2GB, 850MHz
  • GPU device 1: AMD/ATI Radeon HD 6970 2GB, 880MHz
  • CPU: AMD Phenom II X6 @ 2.8GHz 1055T
  • RAM: 12GB
  • OS: OpenSUSE 13.1
  • SDK: AMD APP SDK 2.9
Node 2
  • GPU device 0: AMD/ATI Radeon HD 7970 3GB, 1000MHz
  • GPU device 1: nVidia GeForce GTX 560 Ti 2GB, 822MHz
  • CPU: Intel Core i7-2600 @ 3.4GHz
  • RAM: 16GB
  • OS: OpenSUSE 12.2
  • SDK: nVidia CUDA Toolkit 6.0.1, AMD APP SDK 2.9

Completed OpenCL project should be uploaded via User dashboard (see instructions and example there), compilation and execution terminal output logs will be provided to the user.

The information send to hgpu.org will be treated according to our Privacy Policy

HGPU group © 2010-2014 hgpu.org

All rights belong to the respective authors

Contact us: