Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Optimization of GPU-based sparse matrix multiplication for large sparse networks

Full metadata record
DC Field Value Language
dc.contributor.authorLee, Jeongmyung-
dc.contributor.authorKang, Seokwon-
dc.contributor.authorYu, Yongseung-
dc.contributor.authorJo, Yong-Yeon-
dc.contributor.authorKim, Sang-Wook-
dc.contributor.authorPark, Yongjun-
dc.date.accessioned2022-07-08T06:09:16Z-
dc.date.available2022-07-08T06:09:16Z-
dc.date.created2021-05-13-
dc.date.issued2020-04-
dc.identifier.issn1084-4627-
dc.identifier.urihttps://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/145874-
dc.description.abstractSparse matrix multiplication (spGEMM) is widely used to analyze the sparse network data, and extract important information based on matrix representation. As it contains a high degree of data parallelism, many efficient implementations using data-parallel programming platforms such as CUDA and OpenCL have been introduced on graphic processing units (GPUs). Several well-known spGEMM techniques, such as cuS- PARSE and CUSP, often do not utilize the GPU resources fully, owing to the load imbalance between threads in the expansion process and high memory contention in the merge process. Furthermore, even though several outer-product-based spGEMM techniques are proposed to solve the load balancing problem on expansion, they still do not utilize the GPU resources fully, because severe computation load variations exist among the multiple thread blocks.To solve these challenges, this paper proposes a new optimization pass called Block Reorganizer, which balances the total computations of each computing unit on target GPUs, based on the outer-product-based expansion process, and reduces the memory pressure during the merge process. For expansion, it first identifies the actual computation amount for each block, and then performs two thread block transformation processes based on their characteristics: 1) B-Splitting to transform a heavy-computation blocks into multiple small blocks and 2) B- Gathering to aggregate multiple small-computation blocks to a larger block. While merging, it improves the overall performance by performing B-Limiting to limit the number of blocks on each computing unit. Experimental results show that it improves the total performance of kernel execution by 1.43x, on an average, when compared to the row-product-based spGEMM, for NVIDIA Titan Xp GPUs on real-world datasets.-
dc.language영어-
dc.language.isoen-
dc.publisherIEEE Computer Society-
dc.titleOptimization of GPU-based sparse matrix multiplication for large sparse networks-
dc.typeArticle-
dc.contributor.affiliatedAuthorKim, Sang-Wook-
dc.contributor.affiliatedAuthorPark, Yongjun-
dc.identifier.doi10.1109/ICDE48307.2020.00085-
dc.identifier.scopusid2-s2.0-85085862760-
dc.identifier.bibliographicCitationProceedings - International Conference on Data Engineering, v.2020-April, pp.925 - 936-
dc.relation.isPartOfProceedings - International Conference on Data Engineering-
dc.citation.titleProceedings - International Conference on Data Engineering-
dc.citation.volume2020-April-
dc.citation.startPage925-
dc.citation.endPage936-
dc.type.rimsART-
dc.type.docTypeConference Paper-
dc.description.journalClass1-
dc.description.isOpenAccessN-
dc.description.journalRegisteredClassscopus-
dc.subject.keywordPlusBalancing-
dc.subject.keywordPlusExpansion-
dc.subject.keywordPlusGraphics processing unit-
dc.subject.keywordPlusParallel programming-
dc.subject.keywordPlusProgram processors-
dc.subject.keywordPlusBlock transformations-
dc.subject.keywordPlusComputation blocks-
dc.subject.keywordPlusData-parallel programming-
dc.subject.keywordPlusEfficient implementation-
dc.subject.keywordPlusGraphic processing units (GPUs)-
dc.subject.keywordPlusLoad balancing problem-
dc.subject.keywordPlusMatrix representation-
dc.subject.keywordPlusReal-world datasets-
dc.subject.keywordPlusMatrix algebra-
dc.subject.keywordAuthorGPU-
dc.subject.keywordAuthorLinear algebra-
dc.subject.keywordAuthorSparse matrix multiplication-
dc.subject.keywordAuthorSparse network-
dc.identifier.urlhttps://ieeexplore.ieee.org/document/9101654-
Files in This Item
Go to Link
Appears in
Collections
서울 공과대학 > 서울 컴퓨터소프트웨어학부 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Park, Yong jun photo

Park, Yong jun
서울 공과대학 (서울 컴퓨터소프트웨어학부)
Read more

Altmetrics

Total Views & Downloads

BROWSE