hgpu.org » nVidia GeForce Titan
Real-Time Dedispersion for Fast Radio Transient Surveys, using Auto Tuning on Many-Core Accelerators
Alessio Sclocco, Joeri van Leeuwen, Henri E. Bal, Rob V. van Nieuwpoort
Tags: Astrophysics, ATI, ATI Radeon HD 7970, Instrumentation and Methods for Astrophysics, Intel Xeon Phi, nVidia, nVidia GeForce GTX 680, nVidia GeForce Titan, OpenCL, OpenMP, Package, Tesla K20
January 12, 2016 by hgpu
Guy L. Steele Jr. (Oracle Labs), Jean-Baptiste Tristan
Tags: Computer science, CUDA, Latent Dirichlet allocation, Machine learning, nVidia, nVidia GeForce Titan
May 16, 2015 by hgpu
Recent source codes
* * *
Most viewed papers (last 30 days)
- Architecture-Aware LLM Inference Optimization on AMD Instinct GPUs: A Comprehensive Benchmark and Deployment Study
- AutoKernel: Autonomous GPU Kernel Optimization via Iterative Agent-Driven Search
- LLMQ: Efficient Lower-Precision LLM Training for Consumer GPUs
- CuTeGen: An LLM-Based Agentic Framework for Generation and Optimization of High-Performance GPU Kernels using CuTe
- DRTriton: Large-Scale Synthetic Data Reinforcement Learning for Triton Kernel Generation
- MobileKernelBench: Can LLMs Write Efficient Kernels for Mobile Devices?
- Mixed-precision numerics in scientific applications: survey and perspectives
- Triton-Sanitizer: A Fast and Device-Agnostic Memory Sanitizer for Triton with Rich Diagnostic Context
- SOL-ExecBench: Speed-of-Light Benchmarking for Real-World GPU Kernels Against Hardware Limits
- MegaTrain: Full Precision Training of 100B+ Parameter Large Language Models on a Single GPU
* * *




