hgpu.org » LLM
Avinash Maurya, Jie Ye, M. Mustafa Rafique, Franck Cappello, Bogdan Nicolae
Tags: Computer science, CUDA, LLM, Memory, nVidia, nVidia A100, nVidia H100, Performance
June 23, 2024 by hgpu
Recent source codes
* * *
Most viewed papers (last 30 days)
- Jailbreaking LLM-Controlled Robots
- Over-synchronization in GPU Programs
- Testing GPU Numerics: Finding Numerical Differences Between NVIDIA and AMD GPUs
- Accelerating Drug Discovery in AutoDock-GPU with Tensor Cores
- Mixed-precision finite element kernels and assembly: Rounding error analysis and hardware acceleration
- Using modern C++ to improve CUDA programs
- General-Purpose Computing on Tensor Processors
- Superpipeline: A Universal Approach for Reducing GPU Memory Usage in Large Models
- LLload: An Easy-to-Use HPC Utilization Tool
- Online Energy Optimization in GPUs: A Multi-Armed Bandit Approach
* * *