1 of 1

The Art of Sparsity: Mastering High-Dimensional Tensor Storage

1

Scientific Achievement

New findings from storing sparse tensors: (1) linear organization provides the best balance between storage size and access time; (2) they can be transformed into 2D tensors for efficient storage with compressed sparse row (CSR)/compressed sparse column(CSC); (3) tree-structured organization offer exceptional performance in storing high-dimensional tensors.

Significance and Impact

Sparse tensors find widespread use in applications, such as machine learning, partial differential equations(PDE) solvers, and graphs. These new findings contribute to a nuanced understanding of sparse tensor storage formats, guiding informed choices in practical applications.

Technical Approach

  • We analyzed both time and storage complexity for five sparse tensor organizations, including coordinate(COO), linear address(LINEAR), general CSC(GCSC), general CSR (GCSR), and Compressed Sparse Fibers (CSF)
  • We designed experiments to evaluate these five organizations with three representative patterns: tridiagonal sparse pattern (TSP), general graph sparse pattern (GSP), and mixed sparse pattern (MSP)

PI(s)/Facility Lead(s): Bin Dong

Collaborating Institutions: Suren Byna (Ohio State/LBNL), Kesheng Wu (LBNL)

ASCR Program: SciDAC RAPIDS2

ASCR PM: Kalyan Perumalla (SciDAC RAPIDS2), Steve Lee (FASTMath)

Publication(s) for this work: B. Dong, S. Byna, K. Wu , et al., “The Art of Sparsity: Mastering High-Dimensional Tensor Storage (Regular Paper),” ESSA 2024: 5th Workshop on Extreme-Scale Storage and Analysis in conjunction with IEEE IPDPS 2024, San Francisco

Writing Time

Storage Size

Reading Time

the lower

the better

Analysis of both time and storage complexity for five sparse tensor organizations, including coordinate(COO), linear address(LINEAR), general CSC(GCSC), general CSR (GCSR), and Compressed Sparse Fibers (CSF).