Thu Feb 4 06:48:14 PST 2016
- Previous message: [Slony1-general] Cannot fully drop slony node
- Next message: [Slony1-general] Cannot fully drop slony node
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
yes... that's it!! On Thu, Feb 4, 2016 at 8:58 AM, Tignor, Tom <ttignor at akamai.com> wrote: > > If I’m reading right, did you run the drop node op at some point on node 1 > and see it succeed? If it did, the sl_node table on each other node in the > cluster (save perhaps node 3) should show it gone. > If that’s the case, your cluster is fine and you can just run ‘DROP SCHEMA > mycluster CASCADE’ on node 3 and then retry your store node script. > > Tom :-) > > > From: Sung Hsin Lei <sungh.lei at gmail.com> > Date: Wednesday, February 3, 2016 at 11:37 PM > To: slony <slony1-general at lists.slony.info> > Subject: [Slony1-general] Cannot fully drop slony node > > Hey guys, > > I have a cluster with 3 nodes. On the main db, I run the following script: > > > cluster name = slony_cluster; > > node 1 admin conninfo = 'dbname = dbNAME host = localhost user = slonyuser > password = slonPASS port = 5432'; > node 3 admin conninfo = 'dbname = dbNAME host = 172.16.10.4 user = > slonyuser password = slonPASS port = 5432'; > > DROP NODE ( ID = 3, EVENT NODE = 1 ); > > > > I open pdadmin on the main db and I don't see node 3 anymore. However, > when I open pgadmin on the replicated db, I still see node 3. The > replicated db is the one associated with node 3. I run the above script > again on the replicated db but get the following error: > > > C:\Program Files\PostgreSQL\9.3\bin>slonik drop.txt > debug: waiting for 3,5000000004 on 1 > drop.txt:4: PGRES_FATAL_ERROR lock table > "_slony_securithor2".sl_event_lock, "_s > lony_cluster".sl_config_lock;select > "_slony_securithor2".dropNode(ARRAY[3]); > - ERROR: Slony-I: DROP_NODE cannot initiate on the dropped node > > > Now I need to setup another node which must have id=3. I run a script on > the main db(the one pgadmin does not show a node 3). The following is the > script that I used to setup the node and the error that I get: > > > cluster name = slony_cluster; > > node 1 admin conninfo = 'dbname = dbNAME host = localhost user = slonyuser > password = slonPASS port = 5432'; > node 3 admin conninfo = 'dbname = dbNAME host = 172.16.10.4 user = > slonyuser password = slonPASS port = 5432'; > > store node (id=3, comment = 'Slave node 3', event node=1); > store path (server = 1, client = 3, > conninfo='dbname=dbNAME host=172.16.10.3 user=slonyuser password = slonPASS > port = 5432'); > store path (server = 3, client = 1, > conninfo='dbname=dbNAME host=172.16.10.4 user=slonyuser password = slonPASS > port = 5432'); > > subscribe set ( id = 1, provider = 1, receiver = 3, forward = no); > > > > > > C:\Program Files\PostgreSQL\9.3\bin>slonik create.txt > drop.txt:6: Error: namespace "_slony_cluster" already exists in database of > node 3 > > > > Is there another way to drop nodes? Can I recover from this without > dropping the cluster and restarting from scratch? > > > Thanks. > -------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.slony.info/pipermail/slony1-general/attachments/20160204/f335c6ad/attachment.htm
- Previous message: [Slony1-general] Cannot fully drop slony node
- Next message: [Slony1-general] Cannot fully drop slony node
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Slony1-general mailing list