cublas
Here are 81 public repositories matching this topic...
A neat C++ custom Matrix class to perform super-fast GPU (or CPU) powered Matrix/Vector computations with minimal code, leveraging the power of cuBLAS where applicable.
-
Updated
Jun 24, 2017 - C++
GPGPU Inverse Distance Weighting using matrix vector multiplication
-
Updated
Dec 5, 2017 - Cuda
Generalized Orthogonal Least-Squares in CUDA
-
Updated
Apr 21, 2018 - Cuda
Level 3 matrix multiplication using both cublas and mkl.
-
Updated
Jul 20, 2018 - Cuda
🐮 Harness the power of the GPU with '((((((cuBLAS in Common Lisp
-
Updated
Feb 18, 2019 - Common Lisp
Escoin: Efficient Sparse Convolutional Neural Network Inference on GPUs
-
Updated
Feb 28, 2019 - C++
The repository targets the OpenCL gemm function performance optimization. It compares several libraries clBLAS, clBLAST, MIOpenGemm, Intel MKL(CPU) and cuBLAS(CUDA) on different matrix sizes/vendor's hardwares/OS. Out-of-the-box easy as MSVC, MinGW, Linux(CentOS) x86_64 binary provided. 在不同矩阵大小/硬件/操作系统下比较几个BLAS库的sgemm函数性能,提供binary,开盒即用。
-
Updated
Mar 28, 2019 - C
Real-time GPU Beamformer for DSA110 written in C/CUDA
-
Updated
May 21, 2019 - Jupyter Notebook
Improve this page
Add a description, image, and links to the cublas topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the cublas topic, visit your repo's landing page and select "manage topics."