The Art of Sparsity: Mastering High-Dimensional Tensor Storage
1
Scientific Achievement
New findings from storing sparse tensors: (1) linear organization provides the best balance between storage size and access time; (2) they can be transformed into 2D tensors for efficient storage with compressed sparse row (CSR)/compressed sparse column(CSC); (3) tree-structured organization offer exceptional performance in storing high-dimensional tensors.
Significance and Impact
Sparse tensors find widespread use in applications, such as machine learning, partial differential equations(PDE) solvers, and graphs. These new findings contribute to a nuanced understanding of sparse tensor storage formats, guiding informed choices in practical applications.
Technical Approach
We analyzed both time and storage complexity for five sparse tensor organizations, including coordinate(COO), linear address(LINEAR), general CSC(GCSC), general CSR (GCSR), and Compressed Sparse Fibers (CSF)
We designed experiments to evaluate these five organizations with three representative patterns: tridiagonal sparse pattern (TSP), general graph sparse pattern (GSP), and mixed sparse pattern (MSP)
PI(s)/Facility Lead(s): Bin Dong
Collaborating Institutions: Suren Byna (Ohio State/LBNL), Kesheng Wu (LBNL)
ASCR Program: SciDAC RAPIDS2
ASCR PM: Kalyan Perumalla (SciDAC RAPIDS2), Steve Lee (FASTMath)
Publication(s) for this work: B. Dong, S. Byna, K. Wu , et al., “The Art of Sparsity: Mastering High-Dimensional Tensor Storage (Regular Paper),” ESSA 2024: 5th Workshop on Extreme-Scale Storage and Analysis in conjunction with IEEE IPDPS 2024, San Francisco
Writing Time
Storage Size
Reading Time
the lower
the better
Analysis of both time and storage complexity for five sparse tensor organizations, including coordinate(COO), linear address(LINEAR), general CSC(GCSC), general CSR (GCSR), and Compressed Sparse Fibers (CSF).