Re: hundreds of millions row dBs
| От | Pierre-Frédéric Caillaud |
|---|---|
| Тема | Re: hundreds of millions row dBs |
| Дата | |
| Msg-id | opsj3sk0f7cq72hf@musicbox обсуждение исходный текст |
| Ответ на | Re: hundreds of millions row dBs (Tom Lane <tgl@sss.pgh.pa.us>) |
| Список | pgsql-general |
To speed up load :
- make less checkpoints (tweak checkpoint interval and other parameters
in config)
- disable fsync (not sure if it really helps)
- have source data, database tables, and log on three physically
different disks
- have the temporary on a different disk too, or in ramdisk
- gunzip while restoring to read less data from the disk
> "Dann Corbit" <DCorbit@connx.com> writes:
>> Here is an instance where a really big ram disk might be handy.
>> You could create a database on a big ram disk and load it, then build
>> the indexes.
>> Then shut down the database and move it to hard disk.
>
> Actually, if you have a RAM disk, just change the
> $PGDATA/base/nnn/pgsql_tmp
> subdirectory into a symlink to some temp directory on the RAM disk.
> Should get you pretty much all the win with no need to move stuff around
> afterwards.
>
> You have to be sure the RAM disk is bigger than your biggest index
> though.
>
> regards, tom lane
>
> ---------------------------(end of broadcast)---------------------------
> TIP 5: Have you checked our extensive FAQ?
>
> http://www.postgresql.org/docs/faqs/FAQ.html
>
В списке pgsql-general по дате отправления: