Re: a faster compression algorithm for pg_dump

Поиск
Список
Период
Сортировка
От Dimitri Fontaine
Тема Re: a faster compression algorithm for pg_dump
Дата
Msg-id 87iq7uv3f2.fsf@hi-media-techno.com
обсуждение исходный текст
Ответ на Re: a faster compression algorithm for pg_dump  (Tom Lane <tgl@sss.pgh.pa.us>)
Ответы Re: a faster compression algorithm for pg_dump
Список pgsql-hackers
Tom Lane <tgl@sss.pgh.pa.us> writes:
> Well, what we *really* need is a convincing argument that it's worth
> taking some risk for.  I find that not obvious.  You can pipe the output
> of pg_dump into your-choice-of-compressor, for example, and that gets
> you the ability to spread the work across multiple CPUs in addition to
> eliminating legal risk to the PG project.

Well, I like -Fc and playing with the catalog to restore in staging
environments only the "interesting" data. I even automated all the
catalog mangling in pg_staging so that I just have to setup which
schema I want, with only the DDL or with the DATA too.
 The fun is when you want to exclude functions that are used in triggers based on the schema where the function lives,
notthe trigger, BTW, but that's another story. 

So yes having both -Fc and another compression facility than plain gzip
would be good news. And benefiting from a better compression in TOAST
would be good too I guess (small size hit, lots faster, would fit).

Summary : my convincing argument is using the dumps for efficiently
preparing development and testing environments from production data,
thanks to -Fc. That includes skipping data to restore.

Regards,
--
dim


В списке pgsql-hackers по дате отправления:

Предыдущее
От: Dimitri Fontaine
Дата:
Сообщение: Re: testing HS/SR - 1 vs 2 performance
Следующее
От: Simon Riggs
Дата:
Сообщение: Re: Hot Standby: Startup at shutdown checkpoint