Re: Auto-clustering?
От | Pierre C |
---|---|
Тема | Re: Auto-clustering? |
Дата | |
Msg-id | op.vnusbnnqeorkce@apollo13 обсуждение исходный текст |
Ответ на | Auto-clustering? (Royce Ausburn <royce@inomial.com>) |
Ответы |
Re: Auto-clustering?
|
Список | pgsql-performance |
> fc=# explain analyse select collection, period, tariff, sum(bytesSent), > sum(bytesReceived), sum(packets), max(sample), (starttime / 3600) * 3600 > as startchunk from sample_20101001 where starttime between 1287493200 > and 1290171599 and collection=128 and ip = '10.9.125.207' group by > startchunk, tariff, collection, > period; If CLUSTER locks bother you, and you don't do UPDATEs, you might consider doing something like this : - accumulate the rows in a "recent" table - every hour, INSERT INTO archive SELECT * FROM recent ORDER BY (your cluster fields) - DELETE FROM recent the rows you just inserted - VACUUM recent The cluster in your archive table will not be perfect but at least all rows from 1 source in 1 hour will be stored close together. But clustering doesn't need to be perfect either, if you get 100x better locality, that's already good ! Now, if you have a huge amount of data but never query it with a precision exceeding 1 hour, you might consider creating an aggregate table where, at the end of every hour, you only store sum(), min(), max() of the data for the last hour's data using GROUP BY the fields you want. You could also use a trigger, but that would generate a huge amount of UPDATEs. For the above query you'd do : INSERT INTO stats_by_hour (columns...) SELECT collection, ip, period, tariff, sum(bytesSent), sum(bytesReceived), sum(packets), max(sample), (starttime / 3600) * 3600 as startchunk from sample_20101001 WHERE starttime > some value GROUP BY collection, ip, period, tariff, startchunk Then you can run aggregates against this much smaller table instead.
В списке pgsql-performance по дате отправления: