Re: 8K recordsize bad on ZFS?

Поиск
Список
Период
Сортировка
От Josh Berkus
Тема Re: 8K recordsize bad on ZFS?
Дата
Msg-id 4BE9FDFC.40401@agliodbs.com
обсуждение исходный текст
Ответ на 8K recordsize bad on ZFS?  (Josh Berkus <josh@agliodbs.com>)
Список pgsql-performance
> Sure, but bulk load + reandom selects is going to *guarentee*
> fragmentatioon on a COW system (like ZFS, BTRFS, etc) as the selects
> start to write out all the hint-bit-dirtied blocks in random orders...
>
> i.e. it doesn't take long to make an originally nicely continuous block
> random....

I'm testing with DD and Bonnie++, though, which create their own files.

For that matter, running an ETL procedure with a newly created database
on both recordsizes was notably (2.5x) faster on the 128K system.

So I don't think fragmentation is the difference.

--
                                  -- Josh Berkus
                                     PostgreSQL Experts Inc.
                                     http://www.pgexperts.com

В списке pgsql-performance по дате отправления:

Предыдущее
От: Shrirang Chitnis
Дата:
Сообщение: Re: Performance issues when the number of records are around 10 Million
Следующее
От: Josh Berkus
Дата:
Сообщение: Re: Performance issues when the number of records are around 10 Million