I'm making a bit of a guess here, but I suspect the issue is that a
single large dump will hold a transaction open for the entire time. That
will affect vacuums at a minimum; not sure what else could be affected.
On Tue, Nov 22, 2005 at 05:13:44PM -0800, Patrick Hatcher wrote:
>
> OS - RH3
> Pg - 7.4.9
> Ram - 8G
> Disk-709G Raid 0+1
>
> We are having a pgdump issue that we can't seem to find an answer for
>
> Background:
> Production server contains 11 databases of which 1 database comprises 85%
> of the 194G used on the drive. This one large db contains 12 schemas.
> Within the schemas of the large db, there maybe 1 or 2 views that span
> across 2 schemas.
>
> If we do a backup using pgdump against the entire database, it will take
> upwards of 8+ hours for the backup to complete.
>
> If we split the backup up to do a pgdump for the first 10 dbs and then do a
> pgdump by schema on the 1 large db, the the backup takes only 3.5hrs
>
> The other than using the schema switch, there is no compression happening
> on either dump.
>
> Any ideas why this might be happening or where we can check for issues?
>
> TIA
> Patrick Hatcher
> Development Manager Analytics/MIO
> Macys.com
>
>
> ---------------------------(end of broadcast)---------------------------
> TIP 4: Have you searched our list archives?
>
> http://archives.postgresql.org
>
--
Jim C. Nasby, Sr. Engineering Consultant jnasby@pervasive.com
Pervasive Software http://pervasive.com work: 512-231-6117
vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461