Tags: Computer science, CUDA, Metaheuristics, nVidia, nVidia GeForce GTX 1060, Package, Security, Tutorial
Tags: Computer science, Computer vision, Evolutionary Computations, Image registration, Metaheuristics, Neural and Evolutionary Computing, Overview
Tags: Algorithms, Artificial intelligence, Benchmarking, Computer science, CUDA, Metaheuristics, nVidia, nVidia GeForce GTS 780 Ti
Tags: Computer science, Metaheuristics, Neural and Evolutionary Computing, nVidia, nVidia GeForce GTX 960, nVidia GeForce GTX 980, OpenACC, OpenMP, Optimization
Tags: Ant colony optimization, Computer science, CUDA, Metaheuristics, nVidia, nVidia GeForce GTX 680
Tags: Algorithms, Clustering, Computer science, CUDA, Metaheuristics, nVidia, Tesla K40
Tags: Algorithms, Cellular automata, Metaheuristics, nVidia, nVidia GeForce GTX 660 Ti, OpenCL, Package, Signal processing, Thesis
Tags: Algorithms, Computer science, Metaheuristics, nVidia, nVidia GeForce GT 525 M, OpenCL, Optimization, Search
Tags: Algorithms, Biology, Computational biology, Computer science, Heterogeneous systems, Metaheuristics, nVidia, nVidia GeForce GTX 560 Ti, OpenCL, Optimization, Pattern recognition, Thesis
Tags: Computer science, CUDA, Metaheuristics, nVidia, nVidia GeForce GT 630 M, nVidia GeForce GTS 450, Pattern recognition, Thesis
Recent source codes
Most viewed papers (last 30 days)
- SYCL-Bench 2020: Benchmarking SYCL 2020 on AMD, Intel, and NVIDIA GPUs
- Green AI: A Preliminary Empirical Study on Energy Consumption in DL Models Across Different Runtime Infrastructures
- Parallel programming in mobile devices with FancyJCL
- Spyx: A Library for Just-In-Time Compiled Optimization of Spiking Neural Networks
- Benchmarking and Dissecting the Nvidia Hopper GPU Architecture
- Using AI libraries for Incompressible Computational Fluid Dynamics
- FTTN: Feature-Targeted Testing for Numerical Properties of NVIDIA & AMD Matrix Accelerators
- Sustainable Supercomputing for AI: GPU Power Capping at HPC Scale
- APPy: Annotated Parallelism for Python on GPUs
- QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference