TY - GEN
T1 - Accelerating Relative-error Bounded Lossy Compression for HPC datasets with Precomputation-Based Mechanisms
AU - Zou, Xiangyu
AU - Lu, Tao
AU - Xia, Wen
AU - Wang, Xuan
AU - Zhang, Weizhe
AU - Di, Sheng
AU - Tao, Dingwen
AU - Cappello, Franck
N1 - Funding Information:
We are grateful to the anonymous reviewers for their insightful comments and feedback on this work. This research was partly supported by National Key Research and Development Program of China under Grant 2017YFB0802204, NSFC No.61502190, 61672186, and 61872110, the Open Project Program of Wuhan National Laboratory for Optoelectronics NO. 2018WNLOKF008. This research was also supported by the Exascale Computing Project (ECP), Project Number: 17-SC-20-SC, a collaborative effort of two DOE organizations the Office of Science and the National Nuclear Security Administration, responsible for the planning/preparation of a capable exascale ecosystem, including software, applications, hardware, advanced system engineering and early testbed platforms, to support the nation’s exascale computing imperative. The material was supported by the U.S. Department of Energy, Office of Science, under contract DE-AC02-06CH11357, and also by National Science Foundation under Grant No. 1619253. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the funding agencies.
Publisher Copyright:
© 2019 IEEE.
PY - 2019/5
Y1 - 2019/5
N2 - Scientific simulations in high-performance computing (HPC) environments are producing vast volume of data, which may cause a severe I/O bottleneck at runtime and a huge burden on storage space for post-analysis. Unlike the traditional data reduction schemes (such as deduplication or lossless compression), not only can error-controlled lossy compression significantly reduce the data size but it can also hold the promise to satisfy user demand on error control. Point-wise relative error bounds (i.e., compression errors depends on the data values) are widely used by many scientific applications in the lossy compression, since error control can adapt to the precision in the dataset automatically. Point-wise relative error bounded compression is complicated and time consuming. In this work, we develop efficient precomputation-based mechanisms in the SZ lossy compression framework. Our mechanisms can avoid costly logarithmic transformation and identify quantization factor values via a fast table lookup, greatly accelerating the relative-error bounded compression with excellent compression ratios. In addition, our mechanisms also help reduce traversing operations for Huffman decoding, and thus significantly accelerate the decompression process in SZ. Experiments with four well-known real-world scientific simulation datasets show that our solution can improve the compression rate by about 30% and decompression rate by about 70% in most of cases, making our designed lossy compression strategy the best choice in class in most cases.
AB - Scientific simulations in high-performance computing (HPC) environments are producing vast volume of data, which may cause a severe I/O bottleneck at runtime and a huge burden on storage space for post-analysis. Unlike the traditional data reduction schemes (such as deduplication or lossless compression), not only can error-controlled lossy compression significantly reduce the data size but it can also hold the promise to satisfy user demand on error control. Point-wise relative error bounds (i.e., compression errors depends on the data values) are widely used by many scientific applications in the lossy compression, since error control can adapt to the precision in the dataset automatically. Point-wise relative error bounded compression is complicated and time consuming. In this work, we develop efficient precomputation-based mechanisms in the SZ lossy compression framework. Our mechanisms can avoid costly logarithmic transformation and identify quantization factor values via a fast table lookup, greatly accelerating the relative-error bounded compression with excellent compression ratios. In addition, our mechanisms also help reduce traversing operations for Huffman decoding, and thus significantly accelerate the decompression process in SZ. Experiments with four well-known real-world scientific simulation datasets show that our solution can improve the compression rate by about 30% and decompression rate by about 70% in most of cases, making our designed lossy compression strategy the best choice in class in most cases.
KW - Lossy compression
KW - compression rate
KW - high-performance computing
KW - scientific data
UR - http://www.scopus.com/inward/record.url?scp=85073542487&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85073542487&partnerID=8YFLogxK
U2 - 10.1109/MSST.2019.00-15
DO - 10.1109/MSST.2019.00-15
M3 - Conference contribution
AN - SCOPUS:85073542487
T3 - IEEE Symposium on Mass Storage Systems and Technologies
SP - 65
EP - 78
BT - Proceedings - 2019 35th Symposium on Mass Storage Systems and Technologies, MSST 2019
PB - IEEE Computer Society
T2 - 35th Symposium on Mass Storage Systems and Technologies, MSST 2019
Y2 - 20 May 2019 through 24 May 2019
ER -