
* please run the exact `omping` commands from on all cluster-nodes at the same time and post the output - The first shows whether the latency is acceptable even when much traffic goes over the network, the second one usually spots problems with (missing) multicast-queriers * compare the configs '/etc/corosync/nf' on _all_ nodes - it should contain the same data! Note the complete path - '/etc/corosync/nf' is the file that corosync uses for starting up and providing the necessary syncronization for pmxcfs (proxmox cluster filesystem), which is mounted on '/etc/pve' - '/etc/pve/nf' is syncronised by pmxcfs and copied over when changed, but your pmxcfs is currently not running properly: * check the switch logs for anything related to your issues. * check the journal (`journalctl -r` gives you the complete log in reversed order (newest entries first)) for messages from 'pmxcfs', 'corosync' and potential other problems related to network problems. Things that should help in finding the issue: If more information is required, please let us know. Jun 18 18:03:01 node05 pmxcfs: notice: remove message from non-member 10/91328Ĭan anyone help us debug this without having to reboot all hypervisors? Jun 18 13:48:36 node05 pmxcfs: notice: cpg_send_message retried 100 times Jun 18 13:48:36 node05 pmxcfs: notice: cpg_send_message retry 100 Jun 18 13:48:35 node05 pmxcfs: notice: cpg_send_message retry 90 Jun 18 13:48:34 node05 pmxcfs: notice: cpg_send_message retry 80 Jun 18 13:48:33 node05 pmxcfs: notice: cpg_send_message retry 70 Jun 18 13:48:32 node05 pmxcfs: notice: cpg_send_message retry 60 Jun 18 13:48:31 node05 pmxcfs: notice: cpg_send_message retry 50 Jun 18 13:48:30 node05 pmxcfs: notice: cpg_send_message retry 40 Jun 18 13:48:29 node05 pmxcfs: notice: cpg_send_message retry 30 Jun 18 13:48:28 node05 pmxcfs: notice: cpg_send_message retry 20

Jun 18 13:48:27 node05 pmxcfs: notice: cpg_send_message retry 10

#Ecrisper kiosk showing a red x update
Jun 18 13:48:26 node05 pve-firewall: firewall update time (10.002 seconds)
