Analysis of Several Sparse Formats for Matrices used in Sparse-Matrix Dense-Matrix Multiplication for Machine Learning on GPUs
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim, D. | - |
dc.contributor.author | Kim, J. | - |
dc.date.accessioned | 2023-03-08T05:09:48Z | - |
dc.date.available | 2023-03-08T05:09:48Z | - |
dc.date.issued | 2022-10 | - |
dc.identifier.issn | 2162-1233 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/61181 | - |
dc.description.abstract | Sparse-matrix dense-matrix multiplication (SpMM) receives one sparse matrix and one dense matrix as two inputs, and outputs one dense matrix as a result. It plays a vital role in various fields such as deep neural networks graph neural networks and analysis. CUDA, NVIDIA's parallel computing platform, provides cuSPARSE library to support Basic Linear Algebra Subroutines (BLAS) with sparse matrices such as SpMM. In sparse matrices, zero values can be discarded from storage or computations to accelerate execution. In order to represent only non-zero values in sparse matrices, the cuSPARSE library supports several sparse formats for matrices such as COO (COOrdinate), CSR (Compressed Sparse Row), and CSC (Compressed Sparse Column). In addition, since the 3rd Gen. Tensor Cores with Ampere was introduced, CUDA provides cuSPARSELt library for SpMM whose sparse matrix satisfies a 2:4 sparsity pattern, which is approximately 50% sparsity that can occur in machine learning, etc. In this paper, we compare the cuSPARSE library and the cuSPARSELt library for SpMM, in the case of sparse matrices with a 2:4 sparsity pattern(50% sparsity). Furthermore, we compare the performances of three formats to perform SpMM in the cuSPARSE library, in terms of different sparsity such as 75% sparsity, 87.5% sparsity and 99% sparsity. © 2022 IEEE. | - |
dc.format.extent | 3 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | IEEE Computer Society | - |
dc.title | Analysis of Several Sparse Formats for Matrices used in Sparse-Matrix Dense-Matrix Multiplication for Machine Learning on GPUs | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/ICTC55196.2022.9952814 | - |
dc.identifier.bibliographicCitation | International Conference on ICT Convergence, v.2022-October, pp 629 - 631 | - |
dc.description.isOpenAccess | N | - |
dc.identifier.scopusid | 2-s2.0-85143250704 | - |
dc.citation.endPage | 631 | - |
dc.citation.startPage | 629 | - |
dc.citation.title | International Conference on ICT Convergence | - |
dc.citation.volume | 2022-October | - |
dc.type.docType | Conference Paper | - |
dc.publisher.location | 미국 | - |
dc.subject.keywordAuthor | cuSPARSE | - |
dc.subject.keywordAuthor | cuSPARSELt | - |
dc.subject.keywordAuthor | GPUs | - |
dc.subject.keywordAuthor | Machine Learning | - |
dc.subject.keywordAuthor | Sparse-matrix dense-matrix multiplication | - |
dc.subject.keywordAuthor | SpMM | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
84, Heukseok-ro, Dongjak-gu, Seoul, Republic of Korea (06974)02-820-6194
COPYRIGHT 2019 Chung-Ang University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.