Re: array column and b-tree index allowing only 8191 bytes
От | Celso Pinto |
---|---|
Тема | Re: array column and b-tree index allowing only 8191 bytes |
Дата | |
Msg-id | 1213229533.22568.8.camel@starfish обсуждение исходный текст |
Ответ на | Re: array column and b-tree index allowing only 8191 bytes (Alvaro Herrera <alvherre@commandprompt.com>) |
Ответы |
Re: array column and b-tree index allowing only 8191
bytes
|
Список | pgsql-general |
Hi Alvaro, thanks for the hint. I've since experimented with gin and gist and did a small pgbench custom script test. Recalling from my previous message, the int[] on a row can have a maximum of 5000 values. From here I judged gin to be the best option but inserting is really slow. The test was performed on a small EC2 instance. I raised maintenance_work_mem to 512MB but still inserting 50K rows takes more than an hour. I also tested gist, inserts run quickly but running pgbench with 100 clients, each making 10 selects on a random value contained in the int[] takes the machine load to values such as 88 which is definately a no go. What, if any, would be the recommended options to improve this scenario? Not using intarray? :-) Cheers, Celso On Sáb, 2008-06-07 at 12:38 -0400, Alvaro Herrera wrote: > Celso Pinto wrote: > > > So my questions are: is this at all possible? If so, is is possible to > > increate that maximum size? > > Indexing the arrays themselves is probably pretty useless. Try indexing > the elements, which you can do with the intarray contrib module.
В списке pgsql-general по дате отправления: