hgpu.org » nVidia GeFroce RTX 2080 Ti
Dahua Feng, Zhiming Xu, Rongxiang Wang, Felix Xiaozhu Lin
Tags: AI, Apple M2 Max, Apple M2 Pro, Apple M2 Ultra, Computer science, CUDA, Linear Algebra, LLM, Machine learning, nVidia, nVidia GeForce RTX 4090, nVidia GeFroce RTX 2080 Ti, nVidia Quadro RTX 4000, nVidia RTX A6000, Performance, PyTorch
February 3, 2025 by hgpu
Nina Herrmann, Herbert Kuchen
Tags: Computer science, CUDA, Distributed computing, Heterogeneous systems, nVidia, nVidia GeForce GTX 750 Ti, nVidia GeFroce RTX 2080 Ti, nVidia Quadro K620
January 15, 2023 by hgpu
Recent source codes
* * *
Most viewed papers (last 30 days)
- A Microbenchmark Framework for Performance Evaluation of OpenMP Target Offloading
- pyATF: Constraint-Based Auto-Tuning in Python
- TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators
- WgPy: GPU-accelerated NumPy-like array library for web browsers
- CRIUgpu: Transparent Checkpointing of GPU-Accelerated Workloads
- LLMPerf: GPU Performance Modeling meets Large Language Models
- The Shamrock code: I- Smoothed Particle Hydrodynamics on GPUs
- Concurrent Scheduling of High-Level Parallel Programs on Multi-GPU Systems
- TransCL: An Automatic CUDA-to-OpenCL Programs Transformation Framework
- Can Tensor Cores Benefit Memory-Bound Kernels? (No!)
* * *