Author/Authors :
LawrenceIbarria، نويسنده , , PeterLindstrom، نويسنده , , JarekRossignac ، نويسنده , , rzejSzymczak، نويسنده ,
Abstract :
We present a simple method for compressing very large and regularly sampled scalar elds. Our method is particularly
attractive when the entire data set does not t in memory and when the sampling rate is high relative to the
feature size of the scalar eld in all dimensions. Although we report results for R3 and R4 data sets, the proposed
approach may be applied to higher dimensions. The method is based on the new Lorenzo predictor, introduced
here, which estimates the value of the scalar eld at each sample from the values at processed neighbors. The predicted
values are exact when the n-dimensional scalar eld is an implicit polynomial of degree n1. Surprisingly,
when the residuals (differences between the actual and predicted values) are encoded using arithmetic coding,
the proposed method often outperforms wavelet compression in an L1 sense. The proposed approach may be
used both for lossy and lossless compression and is well suited for out-of-core compression and decompression,
because a trivial implementation, which sweeps through the data set reading it once, requires maintaining only a
small buffer in core memory, whose size barely exceeds a single (n1)-dimensional slice of the data.