[ovs-discuss] Raft issues while removing a node

ramteja tadishetti ramtejatadishetti at gmail.com
Fri Nov 9 00:17:03 UTC 2018


Hi,


I am facing trouble in graceful removal of node in a 3 Node RAFT setup.



Node1 :

$ovn-ctl --db-nb-addr=10.8.49.184 --db-nb-port=6641
--db-nb-create-insecure-remote=yes --db-nb-cluster-local-proto=tcp
--db-nb-cluster-local-port=6643 --db-nb-cluster-local-addr=10.8.49.184
start_nb_ovsdb



$ovn-ctl --db-sb-addr=10.8.49.184 --db-sb-port=6642
--db-sb-create-insecure-remote=yes --db-sb-cluster-local-proto=tcp
--db-sb-cluster-local-port=6644 --db-sb-cluster-local-addr=10.8.49.184
start_sb_ovsdb



$ovn-ctl start_northd --ovn-manage-ovsdb=no --ovn-northd-nb-db="tcp:
10.8.49.184:6641,tcp:10.8.49.181:6641,tcp:10.8.49.173:6641"
--ovn-northd-sb-db="tcp:10.8.49.184:6642,tcp:10.8.49.181:6642 tcp:
10.8.49.173:6642"



Node 2:

$ovn-ctl --db-nb-addr=10.8.49.181 --db-nb-port=6641
--db-nb-create-insecure-remote=yes --db-nb-cluster-local-proto=tcp
--db-nb-cluster-local-port=6643 --db-nb-cluster-local-addr=10.8.49.181
--db-nb-cluster-remote-addr=10.8.49.184 start_nb_ovsdb





$ovn-ctl --db-sb-addr=10.8.49.181 --db-sb-port=6642
--db-sb-create-insecure-remote=yes --db-sb-cluster-local-proto=tcp
--db-sb-cluster-local-port=6644 --db-sb-cluster-local-addr=10.8.49.181
--db-sb-cluster-remote-addr=10.8.49.184 start_sb_ovsdb



Node 3:

$ovn-ctl --db-nb-addr=10.8.49.173 --db-nb-port=6641
--db-nb-create-insecure-remote=yes --db-nb-cluster-local-proto=tcp
--db-nb-cluster-local-port=6643 --db-nb-cluster-local-addr=10.8.49.173
--db-nb-cluster-remote-addr=10.8.49.184 start_nb_ovsdb



$ovn-ctl --db-sb-addr=10.8.49.173 --db-sb-port=6642
--db-sb-create-insecure-remote=yes --db-sb-cluster-local-proto=tcp
--db-sb-cluster-local-port=6644 --db-sb-cluster-local-addr=10.8.49.173
--db-sb-cluster-remote-addr=10.8.49.184 start_sb_ovsdb





Now I forcefully stopped northd on Node 1 (Either by doing ovn-central stop
service or forcefully closing the vm in which node1 is running).

 Now Node 3 was elected as leader from remaining nodes(verified by running

$ovs-appctl -t /var/run/openvswitch/ovnnb_db.ctl cluster/status
OVN_Northbound )

$ovs-appctl -t /var/run/openvswitch/ovnsb_db.ctl cluster/status
OVN_Southbound )



Now i tried to remove node using unixctl on the non-leader node.

$ovs-appctl -t /var/run/openvswitch/ovnsb_db.ctl cluster/kick
OVN_Southbound <Node1>



Ideally it should remove node1 but this removed node 2 itself from the
cluster and making the cluster unavailable.



Could you please let us know, how to do a graceful node removal from
cluster if the node becomes unavailable?


Thanks,
Ramteja
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.openvswitch.org/pipermail/ovs-discuss/attachments/20181108/39382489/attachment.html>


More information about the discuss mailing list