Sparsitute
DOE Mathematical Institute for Sparse Computations (2022-2027)
The Sparsity Opportunity
Modern scientific computing faces a fundamental inefficiency: most large-scale computations involve matrices, tensors, and graphs where the vast majority of elements are zero. Without algorithms designed for sparsity, exascale systems waste enormous compute cycles on zeros—multiplying by zero, storing zeros, communicating zeros. Sparsitute exists to close this gap.
Sparsitute is a DOE-funded Mathematical Multifaceted Integrated Capability Center (MMICC), operational from 2022–2027 under the Advanced Scientific Computing Research (ASCR) program. It brings together leading researchers across national laboratories and universities to develop the mathematical foundations and software for sparse computations.
Three Pillars of Sparsity
The institute’s research agenda is organized around three interconnected pillars—unified by the observation that all such structures can be represented with asymptotically fewer than N² nonzero elements.
graph TB
subgraph Sparsitute["Sparsitute — Sparse Computations in Science & Engineering"]
direction TB
SM["<b>Sparse Matrices</b><br/>Structural & data sparsity<br/>from PDEs, ML/AI, kernel<br/>& covariance matrices"]
ST["<b>Sparse Tensors</b><br/>Multi-way relationships<br/>decomposition, completion<br/>& tensor-structured solvers"]
SN["<b>Sparse Networks</b><br/>Graphs, hypergraphs &<br/>simplicial complexes<br/>topological data analysis"]
end
SM --- XC["Cross-Cutting Themes"]
ST --- XC
SN --- XC
XC --> U1["Unified matrix basis<br/>representations & kernels"]
XC --> U2["Communication lower<br/>bound analysis"]
XC --> U3["Sparse LA for<br/>AI/ML training"]
XC --> U4["New algebraic solvers<br/>& preconditioners"]
style Sparsitute fill:#e8f0fe,stroke:#1a73e8,stroke-width:2px
style XC fill:#fce8e6,stroke:#d93025,stroke-width:2px
style SM fill:#e6f4ea,stroke:#137333
style ST fill:#fef7e0,stroke:#e37400
style SN fill:#f3e8fd,stroke:#7627bb
Sparse Matrix Pillar
Goals include developing unified matrix representations and associated primitives, new algebraic solvers and preconditioners targeting numerical optimization, sparse linear algebra algorithms to accelerate AI/ML training, and communication-efficient algorithms informed by lower-bound analysis.
Sparse Tensor Pillar
Targets new algorithms for tensor decomposition and completion, solvers for tensor-structured equations, and optimized computational kernels underlying scalable sparse tensor algorithms.
Sparse Network Pillar
Advances understanding of problems that do not reduce to matrices and tensors over standard fields, where such reductions cause inefficient runtime or parallelization, where temporal data require specialized approaches, and where model flexibility allows inducing sparsity in ML models.
My Role
As Co-PI, I contribute to communication-avoiding algorithms for sparse solvers, sparse tensor factorization methods, and graph algorithm optimization—connecting multiple pillars through cross-cutting algorithmic themes.
Impact
Sparsitute’s work directly feeds into production software (SuperLU_DIST, ArborX) and enables scientific applications from climate modeling to drug discovery. The institute has already produced new theoretical bounds, practical algorithms achieving 25%+ speedups, and community standards for sparse linear algebra interfaces.
Learn more: sparsitute.lbl.gov