Remove server from Cluster Print

  • cluster, proxmox
  • 0

to remove a node out of the cluster you need to do a few things

please note

 

YOU CAN NOT ADD THE SAME NODE AGAIN AFTER YOU REMOVED IT WITH OUT A FRESH INSTALL

so now that is said , lets start the tutorial

first things first

Remove monitor in Ceph 
Remove Manager is Ceph
Remove OSD disks in Ceph
make sure nothing is linked to the server from ceph anymore

then we can go to step 2

move all vps servers to a other node, the one you like to keep atleast
stop all vps servers on this node , Really all vps servers that are left
make sure HA is disabled, for ALL vps servers on this node
Remove all avaible Mounting points like iso libary or any other mounts FROM this node/vps on this node

from this point on , you CANT undo  anymore

we going start the real business now

begin proxmox info :D

Remove a Cluster Node

Caution Read carefully the procedure before proceeding, as it could not be what you want or need.

Move all virtual machines from the node. Make sure you have no local data or backups you want to keep, or save them accordingly. In the following example we will remove the node hp4 from the cluster.

Log in to a different cluster node (not hp4), and issue a pvecm nodes command to identify the node ID to remove:

hp1# pvecm nodes

Membership information
~~~~~~~~~~~~~~~~~~~~~~
    Nodeid      Votes Name
         1          1 hp1 (local)
         2          1 hp2
         3          1 hp3
         4          1 hp4

At this point you must power off hp4 and make sure that it will not power on again (in the network) as it is.

Important As said above, it is critical to power off the node before removal, and make sure that it will never power on again (in the existing cluster network) as it is. If you power on the node as it is, your cluster will be screwed up and it could be difficult to restore a clean cluster state.

After powering off the node hp4, we can safely remove it from the cluster.

 hp1# pvecm delnode hp4

If the operation succeeds no output is returned, just check the node list again with pvecm nodes or pvecm status. You should see something like:

hp1# pvecm status

Quorum information
~~~~~~~~~~~~~~~~~~
Date:             Mon Apr 20 12:44:28 2015
Quorum provider:  corosync_votequorum
Nodes:            3
Node ID:          0x00000001
Ring ID:          1/8
Quorate:          Yes

Votequorum information
~~~~~~~~~~~~~~~~~~~~~~
Expected votes:   3
Highest expected: 3
Total votes:      3
Quorum:           2
Flags:            Quorate

Membership information
~~~~~~~~~~~~~~~~~~~~~~
    Nodeid      Votes Name
0x00000001          1 192.168.15.90 (local)
0x00000002          1 192.168.15.91
0x00000003          1 192.168.15.92

If, for whatever reason, you want this server to join the same cluster again, you have to

  • reinstall Proxmox VE on it from scratch

  • then join it, as explained in the previous section.

Note After removal of the node, its SSH fingerprint will still reside in the known_hosts of the other nodes. If you receive an SSH error after rejoining a node with the same IP or hostname, run pvecm updatecerts once on the re-added node to update its fingerprint cluster wide.

 

end of proxmox info , now lets the cleaning start

You probaly already noticed that you dit not saw those nodes already in the list when you powered down the node

so lets start with the second part

cd /etc/pve/nodes/
ls
rm {node}

{node} is the node to remove ofcourse

now lets move on

do below command on ALL NODES exept ofcourse the one you removed ( as it should be down ... and need to stay down ) 

systemctl stop pve-ha-crm

now execute followign comamnd on a SINGLE node

rm /etc/pve/ha/manager_status

and then we start up everything again

 

systemctl start pve-ha-crm

and then finaly you also remove the node from the crush map ( ceph )

ceph osd crush rm {node}

after this everything should be cleaned out

and now it just look like nothing is changed and that that server was not there in the first place :D


Was dit antwoord nuttig?

« Terug