Re: a few crazy ideas about hash joins
От | Simon Riggs |
---|---|
Тема | Re: a few crazy ideas about hash joins |
Дата | |
Msg-id | 1238776919.5444.201.camel@ebony.2ndQuadrant обсуждение исходный текст |
Ответ на | a few crazy ideas about hash joins (Robert Haas <robertmhaas@gmail.com>) |
Ответы |
Re: a few crazy ideas about hash joins
|
Список | pgsql-hackers |
On Thu, 2009-04-02 at 22:08 -0400, Robert Haas wrote: > 3. Avoid building the exact same hash table twice in the same query. > This happens more often you'd think. For example, a table may have > two columns creator_id and last_updater_id which both reference person > (id). If you're considering a hash join between paths A and B, you > could conceivably check whether what is essentially a duplicate of B > has already been hashed somewhere within path A. If so, you can reuse > that same hash table at zero startup-cost. This is also interesting because there is potential to save memory through that approach, which allows us to allocate work_mem higher and avoid multi-batch altogether. I would be especially interested in using a shared memory hash table that *all* backends can use - if the table is mostly read-only, as dimension tables often are in data warehouse applications. That would give zero startup cost and significantly reduced memory. -- Simon Riggs www.2ndQuadrant.comPostgreSQL Training, Services and Support
В списке pgsql-hackers по дате отправления: