Обсуждение: streaming replication master can fail to shut down

Поиск
Список
Период
Сортировка

streaming replication master can fail to shut down

От
Nick Cleaton
Дата:
I've found that the master server does not finish shutting down in
response to SIGINT to the postmaster when a streaming replica is
connected, if and only if all of the following hold:

* The master server is a specific new bit of kit, which has slightly
higher performance CPUs than the replica and is much more idle.
* The master is configured with archive_mode=on and a non-empty archive command.
* I do not have "strace -p ..." running on the master against the wal
sender servicing the slow replica.

While it's failing to shut down, about 10 megabytes per second of
network traffic is exchanged between master and replica, and the wal
receiver on the replica is at 100% CPU whereas the wal sender on the
master is around 70% to 80%. The wal receiver has about 4MiB of queued
tcp input, the wal sender has a few bytes.This appears to continue
indefinitely, I've tried up to 30 minutes.

If I "strace -p ... >/dev/null 2>&1" the wal sender during the blocked
shutdown, it seems to slow it down enough (wal sender's cpu usage
drops to about 50%) that after about 10 to 20 seconds the shutdown
completes cleanly and the replica logs "end of wal".

I've seen this only with one specific server acting as the master, so
I cannot absolutely rule out a hardware issue of some kind.

This was originally observed with 9.4.6, but I've now replicated it
with 9.5.1 using the pgdg packages under Ubuntu:

postgresql-9.5                      9.5.1-1.pgdg14.04+1
postgresql-client-9.5             9.5.1-1.pgdg14.04+1
postgresql-contrib-9.5          9.5.1-1.pgdg14.04+1

Under 9.5.1 I've observed this by simply installing the packages and
then following https://wiki.postgresql.org/wiki/Binary_Replication_Tutorial#Binary_Replication_in_7_Steps
with only the addition of archive_mode=on and
archive_command=/bin/true on the master.

CPUs on the master: dual Intel(R) Xeon(R) CPU E5-2697 v3 @ 2.60GHz,
hyperthreading off.

CPUs on the replica: dual Intel(R) Xeon(R) CPU E5-2630 v2 @ 2.60GHz,
hyperthreading off.

Re: streaming replication master can fail to shut down

От
Nick Cleaton
Дата:
On 9 March 2016 at 16:17, Nick Cleaton <nick@cleaton.net> wrote:
> I've found that the master server does not finish shutting down in
> response to SIGINT to the postmaster when a streaming replica is
> connected, if and only if all of the following hold:
>
> * The master server is a specific new bit of kit, which has slightly
> higher performance CPUs than the replica and is much more idle.
> * The master is configured with archive_mode=on and a non-empty archive command.
> * I do not have "strace -p ..." running on the master against the wal
> sender servicing the slow replica.

To debug the replication conversation during a failing master shutdown
with my 9.5.1 test installation, I've now also set ssl=false in the
server config and then:

* start a tcpdump
* start the master
* start the replica
* sleep 60
* kill -INT the postmaster on the master

A libpcap dump of the first 100k packets exchanged is at
http://nick.cleaton.net/protodump-100k-nossl.xz (800k compressed)

Highlights from pgshark: at master shutdown time there's:

Packet: t=1457611872.617496, session=17578864249730
PGSQL: type=XLogData, B -> F
XLOG DATA len=2896

... and everything following that is a storm of:

Packet: t=1457611872.761098, session=17578864249730
PGSQL: type=PrimaryKeepalive, B -> F
PRIMARY KEEP-ALIVE

Packet: t=1457611872.761221, session=17578864249730
PGSQL: type=StandbyStatusUpdate, F -> B
STANDBY STATUS UPDATE

Packet: t=1457611872.761363, session=17578864249730
PGSQL: type=PrimaryKeepalive, B -> F
PRIMARY KEEP-ALIVE

There were a total of 25486 StandbyStatusUpdate messages and 64769
PrimaryKeepalive messages during the 0.6 seconds between the start of
master shutdown and tcpdump hitting the 100k packet limit.


Relevant top and netstat output about 10 minutes into the blocked shutdown:

          PID  USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM
TIME+ COMMAND
master:  6167  postgres  20   0  251632   3020   1372 S  74.7  0.0
0:10.64 postgres: wal sender process postgres 10.122.82.98(49730)
streaming 0/11000098
replica: 28211 postgres  20   0  263100   4580   1964 R 100.0  0.0
0:21.26 postgres: wal receiver process   streaming 0/11000000

         Proto Recv-Q  Send-Q Local Address           Foreign Address
       State
master:  tcp        0 2553544 10.122.83.27:5432
10.122.82.98:49730      ESTABLISHED
replica: tcp  4284291     390 10.122.82.98:49730
10.122.83.27:5432       ESTABLISHED

Re: streaming replication master can fail to shut down

От
Nick Cleaton
Дата:
On 10 March 2016 at 13:13, Nick Cleaton <nick@cleaton.net> wrote:

> ... and everything following that is a storm of:
>
> Packet: t=1457611872.761098, session=17578864249730
> PGSQL: type=PrimaryKeepalive, B -> F
> PRIMARY KEEP-ALIVE
>
> Packet: t=1457611872.761221, session=17578864249730
> PGSQL: type=StandbyStatusUpdate, F -> B
> STANDBY STATUS UPDATE
>
> Packet: t=1457611872.761363, session=17578864249730
> PGSQL: type=PrimaryKeepalive, B -> F
> PRIMARY KEEP-ALIVE

This patch is enough to eliminate the problem on my hardware; in this
test the server sent only 29 keepalives during the shutdown:
http://nick.cleaton.net/protodump-100k-nossl-patched.xz (47k)

Вложения

Re: streaming replication master can fail to shut down

От
Andres Freund
Дата:
Hi,

On 2016-03-11 14:12:37 +0000, Nick Cleaton wrote:
> This patch is enough to eliminate the problem on my hardware

> diff -Nurd postgresql-9.5.1.orig/src/backend/replication/walsender.c
postgresql-9.5.1/src/backend/replication/walsender.c
> --- postgresql-9.5.1.orig/src/backend/replication/walsender.c    2016-02-08 21:12:28.000000000 +0000
> +++ postgresql-9.5.1/src/backend/replication/walsender.c    2016-03-11 11:56:41.121361222 +0000
> @@ -2502,8 +2502,10 @@
>
>          proc_exit(0);
>      }
> -    if (!waiting_for_ping_response)
> +    if (!waiting_for_ping_response) {
>          WalSndKeepalive(true);
> +        waiting_for_ping_response = true;
> +    }
>  }

That looks (besides non-postges paren placement), reasonable. Will
commit & backpatch (to 9.4, where it looks like the bug has been
introduced).


> in this test the server sent only 29 keepalives during the shutdown:
> http://nick.cleaton.net/protodump-100k-nossl-patched.xz (47k)

I guess you have a fair amount of WAL traffic, and the receiver was
behind a good bit?

Greetings,

Andres Freund

Re: streaming replication master can fail to shut down

От
Nick Cleaton
Дата:
On 28 April 2016 at 19:14, Andres Freund <andres@anarazel.de> wrote:
>
>> in this test the server sent only 29 keepalives during the shutdown:
>> http://nick.cleaton.net/protodump-100k-nossl-patched.xz (47k)
>
> I guess you have a fair amount of WAL traffic, and the receiver was
> behind a good bit?

No, IIRC this was on the test cluster that I installed for the purpose
of replicating the problem under 9.5; it was essentially idle.

Re: streaming replication master can fail to shut down

От
Andres Freund
Дата:
On 2016-04-28 20:04:21 +0100, Nick Cleaton wrote:
> On 28 April 2016 at 19:14, Andres Freund <andres@anarazel.de> wrote:
> >
> >> in this test the server sent only 29 keepalives during the shutdown:
> >> http://nick.cleaton.net/protodump-100k-nossl-patched.xz (47k)
> >
> > I guess you have a fair amount of WAL traffic, and the receiver was
> > behind a good bit?
>
> No, IIRC this was on the test cluster that I installed for the purpose
> of replicating the problem under 9.5; it was essentially idle.

The reason I'm asking is that I so far can't really replicate the issue
so far. It's pretty clear that waiting_for_ping_response = true; is
needed, but I'm suspicious that that's not all.

Was your standby on a separate machine? What kind of latency?

Andres

Re: streaming replication master can fail to shut down

От
Nick Cleaton
Дата:
On 29 April 2016 at 04:38, Andres Freund <andres@anarazel.de> wrote:

>> > I guess you have a fair amount of WAL traffic, and the receiver was
>> > behind a good bit?
>>
>> No, IIRC this was on the test cluster that I installed for the purpose
>> of replicating the problem under 9.5; it was essentially idle.
>
> The reason I'm asking is that I so far can't really replicate the issue
> so far. It's pretty clear that waiting_for_ping_response = true; is
> needed, but I'm suspicious that that's not all.
>
> Was your standby on a separate machine?

Yes, I've only seen it happen when the standby was on a machine with
slower CPU cores than the primary. All my attempts to replicate it on
a single machine by trying to slow down the wal receiver have failed.
I'm fairly convinced it's some sort of race that depends on wal sender
+ network being faster than wal receiver.

> What kind of latency?

1G switches.

root@XXX:~# ping XXX
PING XXX) 56(84) bytes of data.
64 bytes from XXX: icmp_seq=1 ttl=64 time=0.162 ms
64 bytes from XXX: icmp_seq=2 ttl=64 time=0.223 ms
64 bytes from XXX: icmp_seq=3 ttl=64 time=0.122 ms
64 bytes from XXX: icmp_seq=4 ttl=64 time=0.126 ms
64 bytes from XXX: icmp_seq=5 ttl=64 time=0.149 ms

Re: streaming replication master can fail to shut down

От
Magnus Hagander
Дата:
On Apr 29, 2016 05:38, "Andres Freund" <andres@anarazel.de> wrote:
>
> On 2016-04-28 20:04:21 +0100, Nick Cleaton wrote:
> > On 28 April 2016 at 19:14, Andres Freund <andres@anarazel.de> wrote:
> > >
> > >> in this test the server sent only 29 keepalives during the shutdown:
> > >> http://nick.cleaton.net/protodump-100k-nossl-patched.xz (47k)
> > >
> > > I guess you have a fair amount of WAL traffic, and the receiver was
> > > behind a good bit?
> >
> > No, IIRC this was on the test cluster that I installed for the purpose
> > of replicating the problem under 9.5; it was essentially idle.
>
> The reason I'm asking is that I so far can't really replicate the issue
> so far. It's pretty clear that waiting_for_ping_response = true; is
> needed, but I'm suspicious that that's not all.
>
> Was your standby on a separate machine? What kind of latency?
>

I had no problem reproducing it on my mostly idle laptop. Not enough to
make a proper slowdown, but if I instrumented the receiver it saw the
multiple packets every single time.

/Magnus

Re: streaming replication master can fail to shut down

От
Andres Freund
Дата:
Hi,

I pushed a fix for this to 9.4,9.5 and master yesterday. I'm not
convinced it's all that needs to be fixed, particularly for Magnus'
report.

On 2016-04-29 08:05:51 +0100, Nick Cleaton wrote:
> On 29 April 2016 at 04:38, Andres Freund <andres@anarazel.de> wrote:
>
> >> > I guess you have a fair amount of WAL traffic, and the receiver was
> >> > behind a good bit?
> >>
> >> No, IIRC this was on the test cluster that I installed for the purpose
> >> of replicating the problem under 9.5; it was essentially idle.
> >
> > The reason I'm asking is that I so far can't really replicate the issue
> > so far. It's pretty clear that waiting_for_ping_response = true; is
> > needed, but I'm suspicious that that's not all.
> >
> > Was your standby on a separate machine?
>
> Yes, I've only seen it happen when the standby was on a machine with
> slower CPU cores than the primary. All my attempts to replicate it on
> a single machine by trying to slow down the wal receiver have failed.
> I'm fairly convinced it's some sort of race that depends on wal sender
> + network being faster than wal receiver.

Yes, that's kind of what I'm expecting. You'll only hit that branch if
there's outstanding data to be replicated, but the message has been
handed to the os (!pq_is_send_pending()). Locally that's just a small
data volume, but over actual network on a longer lived connection that
can be a lot more.

Andres