Luke Lonergan wrote:
> Jim,
>
> On 2/26/06 10:37 AM, "Jim C. Nasby" <jnasby@pervasive.com> wrote:
>
> > So the cutover point (on your system with very fast IO) is 4:1
> > compression (is that 20 or 25%?).
>
> Actually the size of the gzipp'ed binary file on disk was 65MB, compared to
> 177.5MB uncompressed, so the compression ratio is 37% (?), or 2.73:1.
I doubt our algorithm would give the same compression (though I haven't
really measured it). The LZ implementation we use is supposed to have
lightning speed at the cost of a not-so-good compression ratio.
> No, unfortunately not. O'Reilly's jobs data have 65K rows, so that would
> work. How do we implement LZW compression on toasted fields? I've never
> done it!
See src/backend/utils/adt/pg_lzcompress.c
--
Alvaro Herrera http://www.CommandPrompt.com/
The PostgreSQL Company - Command Prompt, Inc.