SparseX: Synergizing GPU Libraries for Sparse Matrix Multiplication on Heterogeneous Processors
This program is tentative and subject to change.
Sparse Matrix-Matrix Multiplication (SpMM) on GPU is critical to applications ranging from scientific simulations to Graph Neural Networks (GNNs) and Deep Neural Networks (DNNs). Modern GPUs offer diverse processing units, such as CUDA cores, Tensor Cores, and Sparse Tensor Cores. Many SpMM libraries have been built to harness those different types of processors. Although impressive performance has been reported by each, including cuSparse, Sputnik, CLASP, and Jigsaw, a systematic study in this work shows that no single library is a clear winner across all matrices and scenarios. Based on the empirical observations, this work proposes the first solution to synergize the various libraries to best harness the heterogeneous processors and the array of cutting-edge libraries. The solution is an extensible framework, namely SparseX, that can automatically select the best matrix multiplication library (and types of processors) on the fly for a given sparse matrix on a GPU through an agile accurate predictive model. Experiments show that SparseX can speed up sparse matrix multiplications on thousands of real-world matrices significantly over the SOTA GPU libraries, achieving significant speedups (e.g., as much as 95.34x over cuSparse). Its extensible design makes it easy to be extended to cover new libraries and hardware architectures.
This program is tentative and subject to change.
Tue 3 FebDisplayed time zone: Hobart change
09:50 - 11:10 | |||
09:50 20mTalk | TPDE: A Fast Adaptable Compiler Back-End Framework Main Conference | ||
10:10 20mTalk | Synthesizing Instruction Selection Back-Ends from ISA Specifications Made Practical Main Conference | ||
10:30 20mTalk | SparseX: Synergizing GPU Libraries for Sparse Matrix Multiplication on Heterogeneous Processors Main Conference Ruifeng Zhang North Carolina State University, Xiangwei Wang North Carolina State University, Ang Li Pacific Northwest National Laboratory, Xipeng Shen North Carolina State University | ||
10:50 20mTalk | Compilation of Generalized Matrix Chains with Symbolic Sizes Main Conference | ||