KBLAS: an optimized library for dense matrix-vector multiplication on GPU accelerators
DOI10.1145/2818311zbMATH Open1369.65042arXiv1410.1726OpenAlexW1839773802WikidataQ113310224 ScholiaQ113310224MaRDI QIDQ5270751FDOQ5270751
Authors: Ahmad Abdelfattah, Hatem Ltaief, D. E. Keyes
Publication date: 30 June 2017
Published in: ACM Transactions on Mathematical Software (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1410.1726
Recommendations
- Accelerating GPU kernels for dense linear algebra
- Redesigning triangular dense matrix computations on GPUs
- Optimizing sparse matrix-matrix multiplication for the GPU
- High-performance matrix-matrix multiplications of very small matrices
- Batched triangular dense linear algebra kernels for very small matrix sizes on GPUs
Complexity and performance of numerical algorithms (65Y20) Packaged methods for numerical algorithms (65Y15) Numerical algorithms for specific classes of architectures (65Y10) Numerical linear algebra (65Fxx)
Cites Work
Cited In (4)
Uses Software
This page was built for publication: KBLAS: an optimized library for dense matrix-vector multiplication on GPU accelerators
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5270751)