Parallel Tensor Compression for Large-Scale Scientific Data
Citations Over TimeTop 10% of 2016 papers
Abstract
As parallel computing trends towards the exascale, scientific data produced by high-fidelity simulations are growing increasingly massive. For instance, a simulation on a three-dimensional spatial grid with 512 points per dimension that tracks 64 variables per grid point for 128 time steps yields 8 TB of data, assuming double precision. By viewing the data as a dense five-way tensor, we can compute a Tucker decomposition to find inherent low-dimensional multilinear structure, achieving compression ratios of up to 5000 on real-world data sets with negligible loss in accuracy. So that we can operate on such massive data, we present the first-ever distributed-memory parallel implementation for the Tucker decomposition, whose key computations correspond to parallel linear algebra operations, albeit with nonstandard data layouts. Our approach specifies a data distribution for tensors that avoids any tensor data redistribution, either locally or in parallel. We provide accompanying analysis of the computation and communication costs of the algorithms. To demonstrate the compression and accuracy of the method, we apply our approach to real-world data sets from combustion science simulations. We also provide detailed performance results, including parallel performance in both weak and strong scaling experiments.
Related Papers
- → Static and Streaming Tucker Decomposition for Dense Tensors(2022)21 cited
- → Multilinear Side-Information based Discriminant Analysis for face and kinship verification in the wild(2018)25 cited
- → Sparse and Low-rank Tucker Decomposition with Its Application to 2D+3D Facial Expression Recognition(2020)4 cited
- Automatic face annotation by multilinear AAM with Missing Values(2012)
- → On the Stability of Multilinear Dynamical Systems(2021)4 cited