TY - JOUR
T1 - Performance Optimization for Relative-Error-Bounded Lossy Compression on Scientific Data
AU - Zou, Xiangyu
AU - Lu, Tao
AU - Xia, Wen
AU - Wang, Xuan
AU - Zhang, Weizhe
AU - Zhang, Haijun
AU - Di, Sheng
AU - Tao, Dingwen
AU - Cappello, Franck
N1 - Funding Information:
The authors would like to thank the anonymous reviewers for their insightful comments and feedback on this work. This research was partly supported by National Key R&D Program of China under Grant No. 2018YFB1003800 and No. 2018YFB1003805, NSFC No. 61972441, No. 61832004, No. 61972112, No. 61672186, and No. 61872110, the Open Project Program of Wuhan National Laboratory for Optoelectronics No. 2018WNLOKF008, Key R&D Program for Guangdong Province under Grant No.2019B010136001, the Shenzhen Science and Technology Program under Grant No. JCYJ20170413105929681 and No. JCYJ20170811161545863. The material was supported by the U.S. Department of Energy, Office of Science, under contract DE-AC02-06CH11357, and also by National Science Foundation under Grant No. 1619253. Any opinions, findings, and conclusions or recommendations expressed in this article are those of the author(s) and do not necessarily reflect the views of the funding agencies. The preliminary version of this article appeared in the proceedings of IEEE MSST 2019. X. Zou and T. Lu equally contributed to this work.
Publisher Copyright:
© 1990-2012 IEEE.
PY - 2020/7/1
Y1 - 2020/7/1
N2 - Scientific simulations in high-performance computing (HPC) environments generate vast volume of data, which may cause a severe I/O bottleneck at runtime and a huge burden on storage space for postanalysis. Unlike traditional data reduction schemes such as deduplication or lossless compression, not only can error-controlled lossy compression significantly reduce the data size but it also holds the promise to satisfy user demand on error control. Pointwise relative error bounds (i.e., compression errors depends on the data values) are widely used by many scientific applications with lossy compression since error control can adapt to the error bound in the dataset automatically. Pointwise relative-error-bounded compression is complicated and time consuming. In this article, we develop efficient precomputation-based mechanisms based on the SZ lossy compression framework. Our mechanisms can avoid costly logarithmic transformation and identify quantization factor values via a fast table lookup, greatly accelerating the relative-error-bounded compression with excellent compression ratios. In addition, we reduce traversing operations for Huffman decoding, significantly accelerating the decompression process in SZ. Experiments with eight well-known real-world scientific simulation datasets show that our solution can improve the compression and decompression rates (i.e., the speed) by about 40 and 80 p, respectively, in most of cases, making our designed lossy compression strategy the best-in-class solution in most cases.
AB - Scientific simulations in high-performance computing (HPC) environments generate vast volume of data, which may cause a severe I/O bottleneck at runtime and a huge burden on storage space for postanalysis. Unlike traditional data reduction schemes such as deduplication or lossless compression, not only can error-controlled lossy compression significantly reduce the data size but it also holds the promise to satisfy user demand on error control. Pointwise relative error bounds (i.e., compression errors depends on the data values) are widely used by many scientific applications with lossy compression since error control can adapt to the error bound in the dataset automatically. Pointwise relative-error-bounded compression is complicated and time consuming. In this article, we develop efficient precomputation-based mechanisms based on the SZ lossy compression framework. Our mechanisms can avoid costly logarithmic transformation and identify quantization factor values via a fast table lookup, greatly accelerating the relative-error-bounded compression with excellent compression ratios. In addition, we reduce traversing operations for Huffman decoding, significantly accelerating the decompression process in SZ. Experiments with eight well-known real-world scientific simulation datasets show that our solution can improve the compression and decompression rates (i.e., the speed) by about 40 and 80 p, respectively, in most of cases, making our designed lossy compression strategy the best-in-class solution in most cases.
KW - Lossy compression
KW - compression rate
KW - high-performance computing
KW - scientific data
UR - http://www.scopus.com/inward/record.url?scp=85082166045&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85082166045&partnerID=8YFLogxK
U2 - 10.1109/TPDS.2020.2972548
DO - 10.1109/TPDS.2020.2972548
M3 - Article
AN - SCOPUS:85082166045
SN - 1045-9219
VL - 31
SP - 1665
EP - 1680
JO - IEEE Transactions on Parallel and Distributed Systems
JF - IEEE Transactions on Parallel and Distributed Systems
IS - 7
M1 - 8989806
ER -