Обсуждение: Benchmarking PGSQL?

Поиск
Список
Период
Сортировка

Benchmarking PGSQL?

От
"Krishna Kumar"
Дата:
Hello All,
I'm a performance engineer, quite interested in getting deep into the PGSQL performance enhancement effort. In that regard, I have the following questions :
1. Is there a benchmarking setup, that I can access online?
2. What benchmarks are we running , for performance numbers?
3. What are the current issues, related to performance?
4. Where can I start, with the PGSQL performance effort?

Thanks a lot,
Krishna

Re: Benchmarking PGSQL?

От
"Shoaib Mir"
Дата:
Have you tried pgbench yet?

--
Shoaib Mir
EnterpriseDB (www.enterprisedb.com)

On 2/14/07, Krishna Kumar <kumar.ramanathan@gmail.com> wrote:
Hello All,
I'm a performance engineer, quite interested in getting deep into the PGSQL performance enhancement effort. In that regard, I have the following questions :
1. Is there a benchmarking setup, that I can access online?
2. What benchmarks are we running , for performance numbers?
3. What are the current issues, related to performance?
4. Where can I start, with the PGSQL performance effort?

Thanks a lot,
Krishna

Re: Benchmarking PGSQL?

От
"Luke Lonergan"
Дата:
Here’s one:

Insert performance is limited to about 10-12 MB/s no matter how fast the underlying I/O hardware.  Bypassing the WAL (write ahead log) only boosts this to perhaps 20 MB/s.  We’ve found that the biggest time consumer in the profile is the collection of routines that “convert to datum”.

You can perform the test using any dataset, you might consider using the TPC-H benchmark kit with a data generator available at www.tpc.org.  Just generate some data, load the schema, then perform some COPY statements, INSERT INTO SELECT FROM and CREATE TABLE AS SELECT.

- Luke


On 2/14/07 2:00 AM, "Krishna Kumar" <kumar.ramanathan@gmail.com> wrote:

Hello All,
I'm a performance engineer, quite interested in getting deep into the PGSQL performance enhancement effort. In that regard, I have the following questions :
1. Is there a benchmarking setup, that I can access online?
2. What benchmarks are we running , for performance numbers?
3. What are the current issues, related to performance?
4. Where can I start, with the PGSQL performance effort?

Thanks a lot,
Krishna


Re: Benchmarking PGSQL?

От
"Merlin Moncure"
Дата:
On 2/14/07, Luke Lonergan <llonergan@greenplum.com> wrote:
>
>  Here's one:
>
>  Insert performance is limited to about 10-12 MB/s no matter how fast the
> underlying I/O hardware.  Bypassing the WAL (write ahead log) only boosts
> this to perhaps 20 MB/s.  We've found that the biggest time consumer in the
> profile is the collection of routines that "convert to datum".
>
>  You can perform the test using any dataset, you might consider using the
> TPC-H benchmark kit with a data generator available at www.tpc.org.  Just
> generate some data, load the schema, then perform some COPY statements,
> INSERT INTO SELECT FROM and CREATE TABLE AS SELECT.

I am curious what is your take on the maximum insert performance, in
mb/sec of large bytea columns (toasted), and how much if any greenplum
was able to advance this over the baseline.  I am asking on behalf of
another interested party.  Interested in numbers broken down per core
on 8 core quad system and also aggreate.

merlin

Re: Benchmarking PGSQL?

От
"Luke Lonergan"
Дата:
Hi Merlin,

On 2/14/07 8:20 AM, "Merlin Moncure" <mmoncure@gmail.com> wrote:

> I am curious what is your take on the maximum insert performance, in
> mb/sec of large bytea columns (toasted), and how much if any greenplum
> was able to advance this over the baseline.  I am asking on behalf of
> another interested party.  Interested in numbers broken down per core
> on 8 core quad system and also aggreate.

Our approach is to attach a segment to each core, so we scale INSERT
linearly on number of cores.  So the per core limit we live with is the
10-20MB/s observed here.  We'd like to improve that so that we get better
performance with smaller machines.

We have demonstrated insert performance of 670 MB/s, 2.4TB/hour for
non-toasted columns using 3 load machines against 120 cores.  This rate was
load machine limited.

WRT toasted bytea columns we haven't done any real benchmarking of those.
Do you have a canned benchmark we can run?

- Luke



Re: Benchmarking PGSQL?

От
Josh Berkus
Дата:
Krisna,

> I'm a performance engineer, quite interested in getting deep into the
> PGSQL performance enhancement effort. In that regard, I have the
> following questions :
> 1. Is there a benchmarking setup, that I can access online?
> 2. What benchmarks are we running , for performance numbers?
> 3. What are the current issues, related to performance?
> 4. Where can I start, with the PGSQL performance effort?

Hey, I work for Sun and we've been working on PostgreSQL & benchmarks.
Hopefully we will soon have a server which runs Spec benchmarks which the
community can legally use for testing (it's waiting on some setup issues).

Help we could use right now includes work on an open source TPCE-like
workload, being run by Rilson and Mark Wong.   Another issue which could
use help is reducing our WAL transaction log volume; it's higher than any
other database.   Or you could work on multi-processor scalability; we are
still trying to identify the bottlenecks which are holding us back there.

--
--Josh

Josh Berkus
PostgreSQL @ Sun
San Francisco

Re: Benchmarking PGSQL?

От
"Merlin Moncure"
Дата:
On 2/14/07, Luke Lonergan <llonergan@greenplum.com> wrote:
> Hi Merlin,
>
> On 2/14/07 8:20 AM, "Merlin Moncure" <mmoncure@gmail.com> wrote:
>
> > I am curious what is your take on the maximum insert performance, in
> > mb/sec of large bytea columns (toasted), and how much if any greenplum
> > was able to advance this over the baseline.  I am asking on behalf of
> > another interested party.  Interested in numbers broken down per core
> > on 8 core quad system and also aggreate.
>
> Our approach is to attach a segment to each core, so we scale INSERT
> linearly on number of cores.  So the per core limit we live with is the
> 10-20MB/s observed here.  We'd like to improve that so that we get better
> performance with smaller machines.
>
> We have demonstrated insert performance of 670 MB/s, 2.4TB/hour for
> non-toasted columns using 3 load machines against 120 cores.  This rate was
> load machine limited.
>
> WRT toasted bytea columns we haven't done any real benchmarking of those.
> Do you have a canned benchmark we can run?

Interested in how fast you can insert binary objects (images, files,
etc). into the database as a file storage system.  Ultimately the
insertions would all be done via libpq ExecPrepared/Params.  A simple
benchmark such as insert a 1mb object via pg_bench over 20 or so
connections would be fine.  Mostly interested in raw throughput per
core and especially interested if you can beat stock pg on the same
hardware.

merlin

Re: Benchmarking PGSQL?

От
"Krishna Kumar"
Дата:
Thanks for the tips, Josh. Would you know of where I can find the TpC-E effort online? I've looked, and have only found references to the 'summer of code' project that Riklas was doing/mentoring.

Also, I'm going to spend some time digging into the code, for the WAL log issue you mentioned. Let's see what I can come up with... (I must confess I'm not a Guru with Database internals, but I'll give this my best...)

Krishna

On 2/15/07, Josh Berkus <josh@agliodbs.com> wrote:
Krisna,

> I'm a performance engineer, quite interested in getting deep into the
> PGSQL performance enhancement effort. In that regard, I have the
> following questions :
> 1. Is there a benchmarking setup, that I can access online?
> 2. What benchmarks are we running , for performance numbers?
> 3. What are the current issues, related to performance?
> 4. Where can I start, with the PGSQL performance effort?

Hey, I work for Sun and we've been working on PostgreSQL & benchmarks.
Hopefully we will soon have a server which runs Spec benchmarks which the
community can legally use for testing (it's waiting on some setup issues).

Help we could use right now includes work on an open source TPCE-like
workload, being run by Rilson and Mark Wong.   Another issue which could
use help is reducing our WAL transaction log volume; it's higher than any
other database.   Or you could work on multi-processor scalability; we are
still trying to identify the bottlenecks which are holding us back there.

--
--Josh

Josh Berkus
PostgreSQL @ Sun
San Francisco