Data Compression for the Exascale Computing Era - Survey

Authors

  • Seung Woo Son Northwestern University, Evanston
  • Zhengzhang Chen Northwestern University, Evanston
  • William Hendrix Northwestern University, Evanston
  • Ankit Agrawal Northwestern University, Evanston
  • Wei-keng Liao Northwestern University, Evanston
  • Alok Choudhary Northwestern University, Evanston

DOI:

https://doi.org/10.14529/jsfi140205

Abstract

While periodic checkpointing has been an important mechanism for tolerating faults in high-performance computing (HPC) systems, it is cost-prohibitive as the HPC system approaches exascale. Applying compression techniques is one common way to mitigate such burdens by reducing the data size, but they are often found to be less effective for scientific datasets. Traditional lossless compression techniques that look for repeated patterns are ineffective for scientific data in which high-precision data is used and hence common patterns are rare to find. In this paper, we present a comparison of several lossless and lossy data compression algorithms and discuss their methodology under the exascale environment. As data volume increases, we discover an increasing trend of new domain-driven algorithms that exploit the inherent characteristics exhibited in many scientific dataset, such as relatively small changes in data values from one simulation iteration to the next or among neighboring data. In particular, significant data reduction has been observed in lossy compression. This paper also discusses how the errors introduced by lossy compressions are controlled and the tradeoffs with the compression ratio.

Downloads

Published

2014-10-01

How to Cite

Son, S. W., Chen, Z., Hendrix, W., Agrawal, A., Liao, W.- keng, & Choudhary, A. (2014). Data Compression for the Exascale Computing Era - Survey. Supercomputing Frontiers and Innovations, 1(2), 76–88. https://doi.org/10.14529/jsfi140205