Issues Adding two servers to an existing Cluster
We have two servers already in a cluster, which was originally configured as a two node cluster. It is in production and has 10 VMs running on each.:
aurora
We have two new server ready to add to a cluster, which has a newer version:
strongest. They do not have any VMs on them yet.
bolivar
At some point along the line I changed the expected to 1, to add the new servers. However, after the change. I see the following when I run the pvecm nodes command.
and I'm not longer able to access strongest/bolivar directly via the web interface. I can access aurora/blooming via the web gui, but when I click on the strongest/bolivar nodes I get the a broken pipe 596 error.
I also reboot strongest, and got the following in the boot log:
I have tested multicasting, and it is enabled on our router.
Does anyone have a way to get all 4 servers on the same cluster, hopefully without rebooting aurora/blooming.
thanks.
myles.
We have two servers already in a cluster, which was originally configured as a two node cluster. It is in production and has 10 VMs running on each.:
aurora
Code:
root@aurora:~# pveversion -v
proxmox-ve-2.6.32: 3.1-109 (running kernel: 2.6.32-23-pve)
pve-manager: 3.1-3 (running version: 3.1-3/dc0e9b0e)
pve-kernel-2.6.32-20-pve: 2.6.32-100
pve-kernel-2.6.32-22-pve: 2.6.32-107
pve-kernel-2.6.32-23-pve: 2.6.32-109
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-1
pve-cluster: 3.0-7
qemu-server: 3.1-1
pve-firmware: 1.0-23
libpve-common-perl: 3.0-6
libpve-access-control: 3.0-6
libpve-storage-perl: 3.0-10
pve-libspice-server1: 0.12.4-1
vncterm: 1.1-4
vzctl: 4.0-1pve3
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-17
ksm-control-daemon: 1.1-1
glusterfs-client: 3.4.0-2
Code:
root@aurora:~# pvecm status
Version: 6.2.0
Config Version: 6
Cluster Name: dii-san-cluster
Cluster Id: 24244
Cluster Member: Yes
Cluster Generation: 408
Membership state: Cluster-Member
Nodes: 2
Expected votes: 2
Total votes: 2
Node votes: 1
Quorum: 2
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: aurora
Node ID: 1
Multicast addresses: 239.192.94.19
Node addresses: 10.10.40.10
Code:
blooming
root@blooming:~# pveversion -v
proxmox-ve-2.6.32: 3.1-109 (running kernel: 2.6.32-23-pve)
pve-manager: 3.1-3 (running version: 3.1-3/dc0e9b0e)
pve-kernel-2.6.32-20-pve: 2.6.32-100
pve-kernel-2.6.32-22-pve: 2.6.32-107
pve-kernel-2.6.32-23-pve: 2.6.32-109
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-1
pve-cluster: 3.0-7
qemu-server: 3.1-1
pve-firmware: 1.0-23
libpve-common-perl: 3.0-6
libpve-access-control: 3.0-6
libpve-storage-perl: 3.0-10
pve-libspice-server1: 0.12.4-1
vncterm: 1.1-4
vzctl: 4.0-1pve3
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-17
ksm-control-daemon: 1.1-1
glusterfs-client: 3.4.0-2
Code:
root@blooming:~# pvecm status
Version: 6.2.0
Config Version: 6
Cluster Name: dii-san-cluster
Cluster Id: 24244
Cluster Member: Yes
Cluster Generation: 408
Membership state: Cluster-Member
Nodes: 2
Expected votes: 2
Total votes: 2
Node votes: 1
Quorum: 2
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: blooming
Node ID: 2
Multicast addresses: 239.192.94.19
Node addresses: 10.10.40.11
strongest. They do not have any VMs on them yet.
Code:
root@strongest:~# pveversion -v
proxmox-ve-2.6.32: 3.1-114 (running kernel: 2.6.32-26-pve)
pve-manager: 3.1-24 (running version: 3.1-24/060bd5a6)
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-2
pve-cluster: 3.0-8
qemu-server: 3.1-8
pve-firmware: 1.0-23
libpve-common-perl: 3.0-9
libpve-access-control: 3.0-8
libpve-storage-perl: 3.0-18
pve-libspice-server1: 0.12.4-2
vncterm: 1.1-6
vzctl: 4.0-1pve4
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-17
ksm-control-daemon: 1.1-1
glusterfs-client: 3.4.1-1
Code:
pvecm status
Version: 6.2.0
Config Version: 6
Cluster Name: dii-san-cluster
Cluster Id: 24244
Cluster Member: Yes
Cluster Generation: 48
Membership state: Cluster-Member
Nodes: 2
Expected votes: 4
Total votes: 2
Node votes: 1
Quorum: 3 Activity blocked
Active subsystems: 1
Flags:
Ports Bound: 0
Node name: strongest
Node ID: 3
Multicast addresses: 239.192.94.19
Node addresses: 10.10.10.78
Code:
root@bolivar:~# pveversion -v
proxmox-ve-2.6.32: 3.1-114 (running kernel: 2.6.32-26-pve)
pve-manager: 3.1-24 (running version: 3.1-24/060bd5a6)
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-2
pve-cluster: 3.0-8
qemu-server: 3.1-8
pve-firmware: 1.0-23
libpve-common-perl: 3.0-9
libpve-access-control: 3.0-8
libpve-storage-perl: 3.0-18
pve-libspice-server1: 0.12.4-2
vncterm: 1.1-6
vzctl: 4.0-1pve4
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-17
ksm-control-daemon: 1.1-1
Code:
root@bolivar:~# pvecm status
Version: 6.2.0
Config Version: 6
Cluster Name: dii-san-cluster
Cluster Id: 24244
Cluster Member: Yes
Cluster Generation: 48
Membership state: Cluster-Member
Nodes: 2
Expected votes: 4
Total votes: 2
Node votes: 1
Quorum: 3 Activity blocked
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: bolivar
Node ID: 4
Multicast addresses: 239.192.94.19
Node addresses: 10.10.10.79
Code:
aurora/blooming
Node Sts Inc Joined Name
1 M 408 2014-02-11 14:57:00 aurora
2 M 396 2014-02-11 14:54:27 blooming
3 X 0 strongest
4 X 0 bolivar
Code:
strongest/bolivar
root@strongest:~# pvecm nodes
Node Sts Inc Joined Name
1 X 0 aurora
2 X 0 blooming
3 M 44 2014-02-11 15:15:40 strongest
4 M 48 2014-02-11 15:16:04 bolivar
I also reboot strongest, and got the following in the boot log:
Code:
Tue Feb 11 15:15:39 2014: Starting cluster:
Tue Feb 11 15:15:39 2014: Checking if cluster has been disabled at boot... [ OK ]
Tue Feb 11 15:15:39 2014: Checking Network Manager... [ OK ]
Tue Feb 11 15:15:39 2014: Global setup... [ OK ]
Tue Feb 11 15:15:39 2014: Loading kernel modules... [ OK ]
Tue Feb 11 15:15:39 2014: Mounting configfs... [ OK ]
Tue Feb 11 15:15:39 2014: Starting cman... [ OK ]
Tue Feb 11 15:15:44 2014: Waiting for quorum... Timed-out waiting for cluster
Tue Feb 11 15:16:28 2014: [FAILED]
Does anyone have a way to get all 4 servers on the same cluster, hopefully without rebooting aurora/blooming.
thanks.
myles.