Tory M Blue tmblue at gmail.com
Sat Sep 10 18:26:21 PDT 2016
On Sat, Sep 10, 2016 at 5:50 PM, Tory M Blue <tmblue at gmail.com> wrote:

>
>
> On Sat, Sep 10, 2016 at 5:44 PM, Tory M Blue <tmblue at gmail.com> wrote:
>
>> Running into an issue, working with a cluster on another network, so I
>> added node 11 (insert secondary in site2) from node 2 (which is the insert
>> secondary of site 1). Set 2 kept failing, so I went in and dropped node 11
>> from the cluster config, but node 2, which is the seecondary insert server
>> continues with this error.. I've not inserted anything into node 2, nor can
>> you, I'm really confused how my primary cluster got in this state..
>>
>>
>> Any assistance to help clean it up? Dropping node 2 and re-adding is a 24
>> hour ordeal.
>>
>> 2016-09-10 17:41:50 PDT ERROR  remoteWorkerThread_11: "insert into
>> "_cls".sl_event     (ev_origin, ev_seqno, ev_timestamp,      ev_snapshot,
>> ev_type     ) values ('11', '5000005698', '2016-09-10 08:34:48.691123-07',
>> '6422065:6530984:6422065', 'SYNC'); insert into "_cls".sl_confirm (con_origin,
>> con_received, con_seqno, con_timestamp)    values (11, 2, '5000005698',
>> now()); select "_cls".logApplySaveStats('_cls', 11, '0.047
>> s'::interval); commit transaction;" PGRES_FATAL_ERROR ERROR:  duplicate key
>> value violates unique constraint "sl_event-pkey"
>>
>> DETAIL:  Key (ev_origin, ev_seqno)=(11, 5000005698) already exists.
>>
>>
> Also interesting node 2 seems to think node 11 is still legit
>
> idb01.prod.ca is node 11:
>
> idb01.prod.ca.domain.net is not active. Make sure that it is already part
> of the slon configuration currently running
>
> select * from _cls.sl_node;
>
>  no_id | no_active | no_comment | no_failed
>
> -------+-----------+------------+-----------
>
>      3 | t         | Node  3    | f
>
>      4 | t         | Node  4    | f
>
>      5 | t         | Node  5    | f
>
>      1 | t         | Node  1    | f
>
>      2 | t         | Node  2    | f
>
>  *   11 | t         | Node  11   | f*
>
>
> Paths are there as well..
>
>
> After manually removing node 11 from all the servers in site 1, node 2
slon started up on it's own and caught up. So not sure what is happening
and why node 11 got into that state

Will look around and make sure that all hosts, know about all hosts

Thanks
Tory
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://lists.slony.info/pipermail/slony1-general/attachments/20160910/26ca5c46/attachment.htm 


More information about the Slony1-general mailing list